<?xml version="1.0" encoding="utf-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.0 20120330//EN" "JATS-journalpublishing1.dtd">
<article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" article-type="research-article">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">INFORMATICA</journal-id>
<journal-title-group><journal-title>Informatica</journal-title></journal-title-group>
<issn pub-type="epub">1822-8844</issn><issn pub-type="ppub">0868-4952</issn><issn-l>0868-4952</issn-l>
<publisher>
<publisher-name>Vilnius University</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="publisher-id">INFOR601</article-id>
<article-id pub-id-type="doi">10.15388/25-INFOR601</article-id>
<article-categories><subj-group subj-group-type="heading">
<subject>Research Article</subject></subj-group></article-categories>
<title-group>
<article-title>Quantifying Binary Classifier Algorithms Similarity with a Consensus Agreement Approach</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author">
<contrib-id contrib-id-type="orcid">https://orcid.org/0000-0001-9180-0270</contrib-id>
<name><surname>Perišić</surname><given-names>Ana</given-names></name><email xlink:href="ana.perisic@pmfst.hr">ana.perisic@pmfst.hr</email><email xlink:href="sisak@vus.hr">sisak@vus.hr</email><xref ref-type="aff" rid="j_infor601_aff_001">1</xref><xref ref-type="aff" rid="j_infor601_aff_002">2</xref><xref ref-type="corresp" rid="cor1">∗</xref><bio>
<p><bold>A. Perišić</bold> received her PhD at the University of Ljubljana (Slovenia) in 2021, after completing her master’s degree in mathematics and postgraduate studies in Economics at University of Zagreb (Croatia). Currently she is working at the Šibenik University of Applied Sciences as a college professor, and at the Department of Mathematics, University of Split as a postdoctoral researcher. Her research interests span a wide range of topics in the development and application of statistical methodologies, with significant contributions to modeling for customer churn prediction, clustering of mixed-type data, developing composite indicators and the development of similarity coefficients for binary data sets. She has contributed to a number of journal articles and conference papers, and participated in diverse research projects, including industry-academia collaborations.</p></bio>
</contrib>
<contrib contrib-type="author">
<contrib-id contrib-id-type="orcid">https://orcid.org/0000-0001-6584-2522</contrib-id>
<name><surname>Vanbelle</surname><given-names>Sophie</given-names></name><email xlink:href="sophie.vanbelle@maastrichtuniversity.nl">sophie.vanbelle@maastrichtuniversity.nl</email><xref ref-type="aff" rid="j_infor601_aff_003">3</xref><bio>
<p><bold>S. Vanbelle</bold> completed a master’s degree in mathematics (ULiègge, Belgium) and a master’s degree in Biostatistics (UHasselt, Belgium) before obtaining her PhD at ULiège in 2009. She is currently associate professor in the department of Methodology &amp; Statistics at the faculty of Health, Medicine and Life Sciences, Maastricht University, The Netherlands. Her research focuses on the development and application of statistical methodology for reliability and agreement studies, with particular interest in complex and multilevel designs and intensive longitudinal data. She has authored numerous peer-reviewed articles and contributed to tutorials and reviews. In addition, she is actively engaged in the statistical community, including service within the Belgian Region of the International Biometric Society.</p></bio>
</contrib>
<contrib contrib-type="author">
<name><surname>Petričević</surname><given-names>Rafaela Brigita</given-names></name><email xlink:href="rafaelap98@gmail.com">rafaelap98@gmail.com</email><xref ref-type="aff" rid="j_infor601_aff_004">4</xref><bio>
<p><bold>R.B. Petričević</bold> graduated in mathematics from the Faculty of Science, University of Split, in 2023. As part of her master’s thesis, she worked on quantifying binary classifier algorithms similarity with a consensus agreement approach. She is currently employed at OTP Bank as a Data Warehouse Specialist, where she works on the implementation of a new data warehouse. Her responsibilities also include developing regulatory reports for the Croatian National Bank and the European Central Bank, as well as supporting data-driven decision-making within the bank.</p></bio>
</contrib>
<aff id="j_infor601_aff_001"><label>1</label>Ruđera Boškovića 33 21000 Split, <institution>University of Split</institution>, <country>Croatia</country></aff>
<aff id="j_infor601_aff_002"><label>2</label>Trg Andrije Hebranga 11, 22000, Šibenik, <institution>Šibenik University of Applied Sciences</institution>, <country>Croatia</country></aff>
<aff id="j_infor601_aff_003"><label>3</label>Peter Debyeplein 1 6229 HA Maastricht, <institution>CAPHRI, Maastricht University</institution>, <country>The Netherlands</country></aff>
<aff id="j_infor601_aff_004"><label>4</label>Split, <institution>Independent Researcher</institution>, <country>Croatia</country></aff>
</contrib-group>
<author-notes>
<corresp id="cor1"><label>∗</label>Corresponding author.</corresp>
</author-notes>
<pub-date pub-type="ppub"><year>2025</year></pub-date><pub-date pub-type="epub"><day>12</day><month>9</month><year>2025</year></pub-date><volume>36</volume><issue>3</issue><fpage>657</fpage><lpage>676</lpage><history><date date-type="received"><month>1</month><year>2025</year></date><date date-type="accepted"><month>9</month><year>2025</year></date></history>
<permissions><copyright-statement>© 2025 Vilnius University</copyright-statement><copyright-year>2025</copyright-year>
<license license-type="open-access" xlink:href="http://creativecommons.org/licenses/by/4.0/">
<license-p>Open access article under the <ext-link ext-link-type="uri" xlink:href="http://creativecommons.org/licenses/by/4.0/">CC BY</ext-link> license.</license-p></license></permissions>
<abstract>
<p>Most classification algorithms involve subjective inputs or hyperparameters to be determined prior to performing the classification. When taking different input or hyperparameter values, each classification algorithm will comprise a collection of classifiers. In this work, we propose a data-driven methodology for assessing similarity in consensus agreement within such a collection of classifiers, and between two classification algorithms, conditional on the dataset of interest. The core of our approach lies in considering the variability introduced by different hyperparameter values for each algorithm when performing such comparisons. We address these problems by evaluating the similarity through consensus agreement and by proposing the application of asymmetric similarity indices based on the Jaccard coefficient. We present the proposed methodology on two publicly available datasets.</p>
</abstract>
<kwd-group>
<label>Key words</label>
<kwd>similarity</kwd>
<kwd>binary classification</kwd>
<kwd>consensus agreement</kwd>
<kwd>Jaccard coefficient</kwd>
<kwd>classifier sets</kwd>
</kwd-group>
</article-meta>
</front>
<body>
<sec id="j_infor601_s_001">
<label>1</label>
<title>Introduction</title>
<p>Classifying objects into classes is a common problem in many research fields. A plethora of supervised machine learning algorithms have been developed: Logistic Regression, Decision Trees, Random Forests (RF), Naive Bayes, Support Vector Machines, Neural Networks and many more. We typically evaluate these algorithms on their predictive performance by comparing predicted and true classes and evaluating measures such as accuracy, sensitivity, and specificity (Han <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor601_ref_007">2012</xref>). This is because we are seeking to find the optimal predictive model, which means finding the model with the highest prediction performance.</p>
<p>Hyperparameter tuning is a crucial step in building an optimal predictive model, as most machine learning algorithms depend on carefully chosen hyperparameter values that significantly impact their performance (Bischl <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor601_ref_002">2023</xref>). For example, when applying RF, the researcher needs to decide on the number of trees and the number of features to sample at each split point. Additionally, other subjective decisions or subjective inputs are required during the model-building process, such as determining the initial set of (relevant) predictors to include. Hyperparameter tuning is usually performed by selecting values that optimize a predetermined metric, such as accuracy, and this process can be automatized (Feurer and Hutter, <xref ref-type="bibr" rid="j_infor601_ref_006">2019</xref>). Other decisions, such as selecting relevant predictors, can be theory-driven and do not necessarily involve hyperparameters, yet can still significantly impact the model and its classification results. The greater the impact of these decisions, the more sensitive the algorithm becomes to hyperparameter values, resulting in an increased diversity of the resulting classifications. This paper focuses on the study of this diversity when applying different hyperparameter values. We propose a data-driven methodology for assessing similarity between two classification algorithms, conditional on the dataset on which the algorithm is applied. This means that classifiers are evaluated through predicted classifications. The novelty of our approach stems from: 
<list>
<list-item id="j_infor601_li_001">
<label>(a)</label>
<p>evaluating each classification algorithm not as a single representative classifier, but as a collection of classifiers built under the same algorithm with varying hyperparameter values,</p>
</list-item>
<list-item id="j_infor601_li_002">
<label>(b)</label>
<p>accounting for the diversity among the corresponding classifiers associated with each binary classification algorithm.</p>
</list-item>
</list> 
The core of this research is therefore in evaluating similarity between two binary classification algorithms when taking into account different hyperparameter values and subjective inputs related to a specific algorithm. Accounting for the diversity of the classifiers associated with each binary classification algorithm imposes two main research questions: 
<list>
<list-item id="j_infor601_li_003">
<label>(R1)</label>
<p>How to evaluate the similarity of classifiers for a given binary classification algorithm when applying different hyperparameter values or different subjective inputs?</p>
</list-item>
<list-item id="j_infor601_li_004">
<label>(R2)</label>
<p>How to evaluate the classification similarity between two binary classification algorithms when applying different hyperparameter values or different subjective inputs for each algorithm?</p>
</list-item>
</list> 
In general, research question (R1) relates to the problem of evaluating similarity between <italic>k</italic> binary classifiers, while research question (R2) can be generalized to the problem of comparing two binary classifier sets. We focus on evaluating the similarity in consensus agreement where we quantify the extent to which all classifiers unanimously predict the same label for a given instance. The formal definition of the consensus agreement is presented in Section <xref rid="j_infor601_s_003">3</xref>. We address these problems by evaluating the similarity through consensus agreement and by proposing the application of asymmetric similarity indices based on the Jaccard coefficient established in Perišić and Vanbelle (<xref ref-type="bibr" rid="j_infor601_ref_017">2024</xref>). Approaches for measuring similarity of binary classifiers are presented in Section <xref rid="j_infor601_s_002">2</xref>, while Section <xref rid="j_infor601_s_003">3</xref> presents the basic theoretical framework related to binary classifiers and consensus agreement, and the methodological framework. In Section <xref rid="j_infor601_s_006">4</xref> we present the application of the proposed framework to publicly available datasets. Section <xref rid="j_infor601_s_009">5</xref> comprises the discussion, conclusion, and propositions for future work.</p>
</sec>
<sec id="j_infor601_s_002">
<label>2</label>
<title>Comparing Binary Classification Algorithms</title>
<p>When measuring similarity, we can have different approaches to what similarity or dissimilarity means. These different approaches can be a result of different definitions of similarity, different fields of application, or different goals of the analysis. Moreover, across different fields, the nomenclature regarding the similarity assessment is not uniform. When evaluating the similarity of two binary classifiers, researchers have used terms (dis)similarity, resemblance, diversity, (dis)agreement. Thus, we are frequently working with the same concept, but using different names.</p>
<p>Evaluating similarity mostly implies quantification of similarity, which is often operationalized by applying a certain similarity measure. The broad application of measuring similarity resulted in a plethora of (dis)simmilarity/agreement/resemblance/diversity measures (see for instance, Choi <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor601_ref_004">2010</xref>). Some of these measures have varying terminology across different fields or their application (for instance, the proportion of agreement and the simple matching coefficient).</p>
<p>When building classification models, especially on imbalanced datasets, we are often more interested in correctly identifying 1s (e.g. presence) than 0s. In such cases, we focus on one class rather than both classes equally. This can be achieved using asymmetric similarity coefficients, the most common being the Jaccard coefficient, which has been used for measuring the agreement between two classifiers. Extensions of this coefficient to multiple classifiers include the <italic>k</italic>-adic Jaccard coefficient (Warrens, <xref ref-type="bibr" rid="j_infor601_ref_029">2009</xref>) and the 2-group <italic>k</italic>-adic Jaccard coefficient (Perišić and Vanbelle, <xref ref-type="bibr" rid="j_infor601_ref_017">2024</xref>).</p>
<p>Many measures of the connection between two classifier outputs can be derived from the statistical literature, but there is less clarity on the subject when three or more classifiers are concerned (Kuncheva and Whitaker, <xref ref-type="bibr" rid="j_infor601_ref_011">2003</xref>). Often, researchers simplify the problem of comparing <inline-formula id="j_infor601_ineq_001"><alternatives><mml:math>
<mml:mi mathvariant="italic">k</mml:mi>
<mml:mo mathvariant="normal">&gt;</mml:mo>
<mml:mn>2</mml:mn></mml:math><tex-math><![CDATA[$k\gt 2$]]></tex-math></alternatives></inline-formula> classifiers by evaluating pairwise similarities where they compare 2 classifiers and then take a weighted linear combination. Also, there have been extensions of some well-known similarity coefficients, like the simple matching and the Jaccard coefficient (Warrens, <xref ref-type="bibr" rid="j_infor601_ref_029">2009</xref>). When it comes to measuring the similarity between two sets of classifiers, the situation worsens. This problem is often reduced to measuring the similarity of two classifiers by applying consensus-based methods. The consensus-based practice determines empirically a consensus category in each group of classifiers and reduces the problem to the case of two classifiers (Vanbelle and Albert, <xref ref-type="bibr" rid="j_infor601_ref_027">2009</xref>). There have been some extensions of the well-known similarity measures, for instance the Cohen’s kappa coefficient (Vanbelle and Albert, <xref ref-type="bibr" rid="j_infor601_ref_027">2009</xref>), the Jaccard coefficient, and the Simple matching coefficient (Perišić and Vanbelle, <xref ref-type="bibr" rid="j_infor601_ref_017">2024</xref>).</p>
<p>Evaluating similarity between binary classifiers can serve different purposes. In addition to comparing technical aspects such as the complexity and execution time of the underlying algorithm, one can focus on comparing their predictive performances or examining their agreement when classifying objects. Researchers have adopted different approaches, focusing on only one or more features for comparison. For instance, Makhtar <italic>et al.</italic> (<xref ref-type="bibr" rid="j_infor601_ref_014">2011</xref>) argue that in order to assess the similarity of predictive models as a whole, it is necessary to assess the similarity of their individual elements (Input, Function and Output) independently. They proposed a methodology to measure classifier similarity based on datasets, model functions, and confusion matrices. This includes the Dataset Similarity Coefficient, Similarity of Predictive Model Function, and Similarity of Output, which are combined into a single overall similarity coefficient.</p>
<p>The agreement between two classifiers can generally be evaluated by comparing the predicted classes for each object, as well as their agreement with reference values. When these reference values are the true classes, the agreement essentially reflects the predictive performance of an algorithm. Rather than comparing the agreement between two classifiers in general, researchers often focus on comparing the agreement between predicted and true classes, which is the basis of the prediction performance assessment, using performance measures like accuracy, sensitivity or AUC (see for instance, Labatut and Cherifi, <xref ref-type="bibr" rid="j_infor601_ref_012">2012</xref>), the Matthews correlation coefficient, the Brier score, Cohen’s Kappa (Chicco <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor601_ref_003">2021</xref>). This is due to the fact that when building a prediction model, we aim to find a classifier that has the highest prediction performance. Given a dataset, the classifier with the highest prediction metrics is selected. As already noted, this procedure can be viewed as assessing the agreement between predicted classes and true classes.</p>
<p>Different procedures have been proposed for comparing multiple classifiers. For instance, the Friedman test (Demšar, <xref ref-type="bibr" rid="j_infor601_ref_005">2006</xref>) followed by tests for multiple comparisons such as the mean-ranks test, sign-test or the Wilcoxon signed-rank test (Benavoli <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor601_ref_001">2016</xref>). A broad range of studies analysed classification metrics in terms of specific characteristics and properties, such as coherency of metrics used for evaluating classifiers, consistency in binary classification, robustness of binary classification performance metrics, etc. Some examples can be found in Shirdel <italic>et al.</italic> (<xref ref-type="bibr" rid="j_infor601_ref_020">2024</xref>).</p>
<p>When assessing the agreement between two classifiers in terms of agreement with true classes, we can further distinguish between correct and incorrect predicted classifications. For instance, Petrakos <italic>et al.</italic> (<xref ref-type="bibr" rid="j_infor601_ref_018">2001</xref>) have separately assessed agreement on correct classification and agreement on incorrect classification by matching the results from individual classifiers. They organize the information as a cross-classification table and then apply the proportion of agreement (the simple matching coefficient), the proportion of specific agreement, and the kappa statistic to assess classifier agreement. Wang <italic>et al.</italic> (<xref ref-type="bibr" rid="j_infor601_ref_028">2023</xref>) established a framework that allows for visual comparison of two classification models and enables the interpretation of the two models’ behaviour discrepancy on different feature conditions. This framework identifies data instances with disagreed predictions from the two compared classifiers and trains a discriminator to learn from the disagreed instances.</p>
<p>Comparing two classifiers by assessing the similarity in their predicted classifications has gained attention in the optimization of ensemble algorithms, often referred to as measuring ensemble diversity. Measuring ensemble diversity, as measuring classifier agreement or similarity, is not straightforward, and as a result, there is no universally accepted formal definition of diversity, nor a universally accepted measure. A natural way to measure the diversity of classifiers is to measure how much their classification decisions differ. Again, this difference can be compared by comparing the predicted classifications, i.e. comparing the predicted class of each object, or by examining the diversity in correct and incorrect classifications. Researchers have proposed a number of ways to measure diversity in ensembles of classifiers (e.g. Kuncheva and Whitaker, <xref ref-type="bibr" rid="j_infor601_ref_011">2003</xref>; Wood <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor601_ref_030">2023</xref>; Tsymbal <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor601_ref_026">2005</xref>) and to generalize a diversity measure between two classifiers to an ensemble of more than two classifiers. For instance, when considering more than two classifiers, we can differentiate between measures averaged over pairs of classifiers and measures defined on the ensemble of classifiers as a whole (Narasimhamurthy, <xref ref-type="bibr" rid="j_infor601_ref_016">2005</xref>). Tsymbal <italic>et al.</italic> (<xref ref-type="bibr" rid="j_infor601_ref_026">2005</xref>) considered different measures of ensemble diversity, where they differentiated between the pairwise (that measure diversity in predictions of a pair of classifiers, or the total ensemble diversity, which is the average of all the classifier pairs in the ensemble), and non-pairwise (which measure diversity in predictions of the whole ensemble only). In the binary classification case, they propose the simple matching coefficient (referred to as the plain disagreement and the fail/non-fail disagreement), the Yule’s Q statistics, the correlation coefficient and the kappa statistic. The non-pairwise measures rely on entropy and ambiguity. Kuncheva and Whitaker (<xref ref-type="bibr" rid="j_infor601_ref_011">2003</xref>) studied ten statistics which can measure diversity among binary classifier outputs (in terms of correct or incorrect vote for the class label): four averaged pairwise measures (the Q statistic, the correlation, the disagreement measure and the double fault measure) and six non-pairwise measures (the entropy of the votes, the difficulty index, the Kohavi-Wolpert variance, the interrater agreement, the generalized diversity, and the coincident failure diversity). All these pairwise measures have been proposed as measures of (dis)similarity in the numerical taxonomy literature (Sokal and Sneath, <xref ref-type="bibr" rid="j_infor601_ref_021">1963</xref>). A theoretical analysis on six existing diversity measures (namely disagreement measure, double fault measure, KW variance, inter-rater agreement, generalized diversity and measure of difficulty) and some underlying relationships between them can be found in Tang <italic>et al.</italic> (<xref ref-type="bibr" rid="j_infor601_ref_025">2006</xref>). Also, one of the most widely used measures of classifier agreement is the Cohen’s kappa coefficient which was also used as a measure of classifier diversity (Margineantu and Dietterich, <xref ref-type="bibr" rid="j_infor601_ref_015">1997</xref>; Zouari <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor601_ref_031">2005</xref>).</p>
</sec>
<sec id="j_infor601_s_003" sec-type="methods">
<label>3</label>
<title>Methodology</title>
<p>This section outlines the proposed methodology for quantifying the similarity between binary classifier algorithms. First, we present the methodological framework for evaluating the similarity between two sets of binary classifiers, as established in Perišić and Vanbelle (<xref ref-type="bibr" rid="j_infor601_ref_017">2024</xref>). Then, we introduce a framework that integrates this approach into a broader methodology for assessing the similarity of binary classifier algorithms.</p>
<sec id="j_infor601_s_004">
<label>3.1</label>
<title>Evaluating Similarity in Consensus Agreement for Binary Classification</title>
<p>We start this section by introducing formal definitions of a binary classifier and a binary classifier set, followed by some fundamental definitions related to classifier agreement. Note that these definitions are made conditionally to a given dataset, i.e. the set of objects to be classified.</p><statement id="j_infor601_stat_001"><label>Definition 1</label>
<title>(<italic>Binary classifier</italic>)<italic>.</italic></title>
<p>Let <inline-formula id="j_infor601_ineq_002"><alternatives><mml:math>
<mml:mi mathvariant="italic">O</mml:mi>
<mml:mo>=</mml:mo>
<mml:mo fence="true" stretchy="false">{</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">o</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo>…</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">o</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo fence="true" stretchy="false">}</mml:mo></mml:math><tex-math><![CDATA[$O=\{{o_{1}},\dots ,{o_{n}}\}$]]></tex-math></alternatives></inline-formula>, <inline-formula id="j_infor601_ineq_003"><alternatives><mml:math>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo stretchy="false">∈</mml:mo>
<mml:mi mathvariant="double-struck">N</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo></mml:math><tex-math><![CDATA[$n\in \mathbb{N},$]]></tex-math></alternatives></inline-formula> be a finite set of <italic>n</italic> objects. A binary classifier on <italic>O</italic> is a function <inline-formula id="j_infor601_ineq_004"><alternatives><mml:math>
<mml:mi mathvariant="italic">c</mml:mi>
<mml:mo>:</mml:mo>
<mml:mi mathvariant="italic">O</mml:mi>
<mml:mo stretchy="false">→</mml:mo>
<mml:mo fence="true" stretchy="false">{</mml:mo>
<mml:mn>0</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo fence="true" stretchy="false">}</mml:mo></mml:math><tex-math><![CDATA[$c:O\to \{0,1\}$]]></tex-math></alternatives></inline-formula>.</p></statement><statement id="j_infor601_stat_002"><label>Definition 2</label>
<title>(<italic>Binary classifier set</italic>)<italic>.</italic></title>
<p>Let <inline-formula id="j_infor601_ineq_005"><alternatives><mml:math>
<mml:mi mathvariant="italic">O</mml:mi>
<mml:mo>=</mml:mo>
<mml:mo fence="true" stretchy="false">{</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">o</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo>…</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">o</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo fence="true" stretchy="false">}</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo></mml:math><tex-math><![CDATA[$O=\{{o_{1}},\dots ,{o_{n}}\},$]]></tex-math></alternatives></inline-formula> <inline-formula id="j_infor601_ineq_006"><alternatives><mml:math>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo stretchy="false">∈</mml:mo>
<mml:mi mathvariant="double-struck">N</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo></mml:math><tex-math><![CDATA[$n\in \mathbb{N},$]]></tex-math></alternatives></inline-formula> be a finite set of <italic>n</italic> objects and <inline-formula id="j_infor601_ineq_007"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">c</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo>…</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">c</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${c_{1}},\dots ,{c_{k}}$]]></tex-math></alternatives></inline-formula>, <inline-formula id="j_infor601_ineq_008"><alternatives><mml:math>
<mml:mi mathvariant="italic">k</mml:mi>
<mml:mo stretchy="false">∈</mml:mo>
<mml:mi mathvariant="double-struck">N</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo></mml:math><tex-math><![CDATA[$k\in \mathbb{N},$]]></tex-math></alternatives></inline-formula> binary classifiers on <italic>O</italic>. Binary classifier set on <italic>O</italic> is a collection of classifiers <inline-formula id="j_infor601_ineq_009"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">c</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo>…</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">c</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${c_{1}},\dots ,{c_{k}}$]]></tex-math></alternatives></inline-formula>. We write <inline-formula id="j_infor601_ineq_010"><alternatives><mml:math>
<mml:mi mathvariant="italic">C</mml:mi>
<mml:mo>=</mml:mo>
<mml:mo fence="true" stretchy="false">{</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">c</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">c</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo>…</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">c</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo fence="true" stretchy="false">}</mml:mo></mml:math><tex-math><![CDATA[$C=\{{c_{1}},{c_{2}},\dots {c_{k}}\}$]]></tex-math></alternatives></inline-formula>.</p></statement>
<p>We are examining the agreement of two binary classifiers. Although it is intuitively clear when two classifiers agree, we present a formal definition.</p><statement id="j_infor601_stat_003"><label>Definition 3.</label>
<p>Let <inline-formula id="j_infor601_ineq_011"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">c</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${c_{1}}$]]></tex-math></alternatives></inline-formula> and <inline-formula id="j_infor601_ineq_012"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">c</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${c_{2}}$]]></tex-math></alternatives></inline-formula> be two binary classifiers. Binary classifiers agree on the classification of an object <italic>o</italic> if <inline-formula id="j_infor601_ineq_013"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">c</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">o</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">c</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">o</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[${c_{1}}(o)={c_{2}}(o)$]]></tex-math></alternatives></inline-formula>.</p></statement>
<p>This corresponds to, depending on the research area, the De Moivre definition of agreement (Hubert, <xref ref-type="bibr" rid="j_infor601_ref_008">1977</xref>), consensus agreement, or the k-adic definition of similarity (Warrens, <xref ref-type="bibr" rid="j_infor601_ref_029">2009</xref>). We evaluate the strength of the agreement on a set of objects <inline-formula id="j_infor601_ineq_014"><alternatives><mml:math>
<mml:mi mathvariant="italic">O</mml:mi>
<mml:mo>=</mml:mo>
<mml:mo fence="true" stretchy="false">{</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">o</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo>…</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">o</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo fence="true" stretchy="false">}</mml:mo></mml:math><tex-math><![CDATA[$O=\{{o_{1}},\dots ,{o_{n}}\}$]]></tex-math></alternatives></inline-formula>. We are interested in the consensus agreement within a binary classifier set and distinguish between two types of consensus agreement: the consensus agreement on the presence <inline-formula id="j_infor601_ineq_015"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">c</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">o</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">c</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">o</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$({c_{1}}(o)={c_{2}}(o)=1)$]]></tex-math></alternatives></inline-formula> and the consensus agreement on the absence <inline-formula id="j_infor601_ineq_016"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">c</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">o</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">c</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">o</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:mn>0</mml:mn>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$({c_{1}}(o)={c_{2}}(o)=0)$]]></tex-math></alternatives></inline-formula>.</p><statement id="j_infor601_stat_004"><label>Definition 4</label>
<title>(<italic>Consensus agreement</italic>)<italic>.</italic></title>
<p>Let <inline-formula id="j_infor601_ineq_017"><alternatives><mml:math>
<mml:mi mathvariant="italic">C</mml:mi>
<mml:mo>=</mml:mo>
<mml:mo fence="true" stretchy="false">{</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">c</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo>…</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">c</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo fence="true" stretchy="false">}</mml:mo></mml:math><tex-math><![CDATA[$C=\{{c_{1}},\dots ,{c_{k}}\}$]]></tex-math></alternatives></inline-formula> be a finite set of classifiers and <inline-formula id="j_infor601_ineq_018"><alternatives><mml:math>
<mml:mi mathvariant="italic">O</mml:mi>
<mml:mo>=</mml:mo>
<mml:mo fence="true" stretchy="false">{</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">o</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo>…</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">o</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo fence="true" stretchy="false">}</mml:mo></mml:math><tex-math><![CDATA[$O=\{{o_{1}},\dots ,{o_{n}}\}$]]></tex-math></alternatives></inline-formula> a finite set of objects. 
<list>
<list-item id="j_infor601_li_005">
<label>(1)</label>
<p>A set of classifiers <italic>C</italic> has a consensus agreement on the presence for an object <inline-formula id="j_infor601_ineq_019"><alternatives><mml:math>
<mml:mi mathvariant="italic">o</mml:mi>
<mml:mo stretchy="false">∈</mml:mo>
<mml:mi mathvariant="italic">O</mml:mi></mml:math><tex-math><![CDATA[$o\in O$]]></tex-math></alternatives></inline-formula> if <inline-formula id="j_infor601_ineq_020"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">c</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">j</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">o</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn></mml:math><tex-math><![CDATA[${c_{j}}(o)=1$]]></tex-math></alternatives></inline-formula>, <inline-formula id="j_infor601_ineq_021"><alternatives><mml:math>
<mml:mi mathvariant="italic">j</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>2</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo>…</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">k</mml:mi></mml:math><tex-math><![CDATA[$j=1,2,\dots ,k$]]></tex-math></alternatives></inline-formula>.</p>
</list-item>
<list-item id="j_infor601_li_006">
<label>(2)</label>
<p>A set of classifiers <italic>C</italic> has a consensus agreement on the absence for an object <inline-formula id="j_infor601_ineq_022"><alternatives><mml:math>
<mml:mi mathvariant="italic">o</mml:mi>
<mml:mo stretchy="false">∈</mml:mo>
<mml:mi mathvariant="italic">O</mml:mi></mml:math><tex-math><![CDATA[$o\in O$]]></tex-math></alternatives></inline-formula> if <inline-formula id="j_infor601_ineq_023"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">c</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">j</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">o</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:mn>0</mml:mn></mml:math><tex-math><![CDATA[${c_{j}}(o)=0$]]></tex-math></alternatives></inline-formula>, <inline-formula id="j_infor601_ineq_024"><alternatives><mml:math>
<mml:mi mathvariant="italic">j</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>2</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo>…</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">k</mml:mi></mml:math><tex-math><![CDATA[$j=1,2,\dots ,k$]]></tex-math></alternatives></inline-formula>.</p>
</list-item>
</list>
</p></statement>
<p>When evaluating the similarity of classifiers <inline-formula id="j_infor601_ineq_025"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">c</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">c</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo>…</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">c</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${c_{1}},{c_{2}},\dots ,{c_{k}}$]]></tex-math></alternatives></inline-formula>, <inline-formula id="j_infor601_ineq_026"><alternatives><mml:math>
<mml:mi mathvariant="italic">k</mml:mi>
<mml:mo>⩾</mml:mo>
<mml:mn>2</mml:mn></mml:math><tex-math><![CDATA[$k\geqslant 2$]]></tex-math></alternatives></inline-formula>, we propose the application of the <inline-formula id="j_infor601_ineq_027"><alternatives><mml:math>
<mml:mi mathvariant="italic">k</mml:mi>
<mml:mo>−</mml:mo></mml:math><tex-math><![CDATA[$k-$]]></tex-math></alternatives></inline-formula>adic Jaccard coefficient, defined as 
<disp-formula id="j_infor601_eq_001">
<label>(1)</label><alternatives><mml:math display="block">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd>
<mml:mi mathvariant="italic">J</mml:mi>
<mml:mo>=</mml:mo><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">a</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">k</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mrow>
</mml:msup>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo>−</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">d</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">k</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mo mathvariant="normal">,</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ J=\frac{{a^{(k)}}}{n-{d^{(k)}}},\]]]></tex-math></alternatives>
</disp-formula> 
where <inline-formula id="j_infor601_ineq_028"><alternatives><mml:math>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">a</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">k</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mrow>
</mml:msup></mml:math><tex-math><![CDATA[${a^{(k)}}$]]></tex-math></alternatives></inline-formula> is the number of objects classified as 1 by all classifiers <inline-formula id="j_infor601_ineq_029"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">c</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">c</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo>…</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">c</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo></mml:math><tex-math><![CDATA[${c_{1}},{c_{2}},\dots ,{c_{k}},$]]></tex-math></alternatives></inline-formula> and <inline-formula id="j_infor601_ineq_030"><alternatives><mml:math>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">d</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">k</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mrow>
</mml:msup></mml:math><tex-math><![CDATA[${d^{(k)}}$]]></tex-math></alternatives></inline-formula> is the number of objects classified as 0 by all classifiers <inline-formula id="j_infor601_ineq_031"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">c</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">c</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo>…</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">c</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>.</mml:mo></mml:math><tex-math><![CDATA[${c_{1}},{c_{2}},\dots ,{c_{k}}.$]]></tex-math></alternatives></inline-formula> The value of this coefficient can be interpreted as the percentage of objects classified as 1 by all classifiers out of objects that were classified as 1 by at least one classifier.</p>
<p>When evaluating the similarity of 2 binary classifier sets, we propose the application of the 2-group k-adic Jaccard coefficient established in Perišić and Vanbelle (<xref ref-type="bibr" rid="j_infor601_ref_017">2024</xref>). The 2-group k-adic Jaccard coefficient is defined as follows. Let <inline-formula id="j_infor601_ineq_032"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">C</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${C_{1}}$]]></tex-math></alternatives></inline-formula> and <inline-formula id="j_infor601_ineq_033"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">C</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${C_{2}}$]]></tex-math></alternatives></inline-formula> be two sets of binary classifiers on the set of objects <italic>O</italic>. The 2-group k-adic Jaccard similarity coefficient is obtained by applying steps (1) to (3) as follows: 
<list>
<list-item id="j_infor601_li_007">
<label>(1)</label>
<p>Determine the k-adic Jaccard coefficient separately for each classifier set by calculating <inline-formula id="j_infor601_ineq_034"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">J</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">C</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">l</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo><mml:mstyle displaystyle="false">
<mml:mfrac>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">a</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">C</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">l</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo>−</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">d</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">C</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">l</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfrac>
</mml:mstyle></mml:math><tex-math><![CDATA[${J_{{C_{l}}}}=\frac{{a_{{C_{l}}}}}{n-{d_{{C_{l}}}}}$]]></tex-math></alternatives></inline-formula>, <inline-formula id="j_infor601_ineq_035"><alternatives><mml:math>
<mml:mi mathvariant="italic">l</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>2</mml:mn></mml:math><tex-math><![CDATA[$l=1,2$]]></tex-math></alternatives></inline-formula>, where <inline-formula id="j_infor601_ineq_036"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">a</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">C</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">l</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${a_{{C_{l}}}}$]]></tex-math></alternatives></inline-formula> is the number of objects classified as 1 by all classifiers from <inline-formula id="j_infor601_ineq_037"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">C</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">l</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${C_{l}}$]]></tex-math></alternatives></inline-formula>, and <inline-formula id="j_infor601_ineq_038"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">d</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">C</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">l</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${d_{{C_{l}}}}$]]></tex-math></alternatives></inline-formula> is the number of objects classified as 0 by all classifiers from <inline-formula id="j_infor601_ineq_039"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">C</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">l</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${C_{l}}$]]></tex-math></alternatives></inline-formula>, <inline-formula id="j_infor601_ineq_040"><alternatives><mml:math>
<mml:mi mathvariant="italic">l</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>2</mml:mn></mml:math><tex-math><![CDATA[$l=1,2$]]></tex-math></alternatives></inline-formula>.</p>
</list-item>
<list-item id="j_infor601_li_008">
<label>(2)</label>
<p>Merge the classifiers from sets <inline-formula id="j_infor601_ineq_041"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">C</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${C_{1}}$]]></tex-math></alternatives></inline-formula> and <inline-formula id="j_infor601_ineq_042"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">C</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${C_{2}}$]]></tex-math></alternatives></inline-formula> into one set <inline-formula id="j_infor601_ineq_043"><alternatives><mml:math>
<mml:mi mathvariant="italic">C</mml:mi>
<mml:mo>=</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">C</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>∪</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">C</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:mo fence="true" stretchy="false">{</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">c</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>11</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo>…</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">c</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">c</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>21</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo>…</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">c</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub>
<mml:mo fence="true" stretchy="false">}</mml:mo></mml:math><tex-math><![CDATA[$C={C_{1}}\cup {C_{2}}=\{{c_{11}},\dots ,{c_{1{k_{1}}}},{c_{21}},\dots ,{c_{2{k_{2}}}}\}$]]></tex-math></alternatives></inline-formula>. Determine the k-adic Jaccard coefficient <inline-formula id="j_infor601_ineq_044"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">J</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">C</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${J_{C}}$]]></tex-math></alternatives></inline-formula> for the classifier set <italic>C</italic>, <inline-formula id="j_infor601_ineq_045"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">J</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">C</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo><mml:mstyle displaystyle="false">
<mml:mfrac>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">a</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">C</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo>−</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">d</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">C</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfrac>
</mml:mstyle></mml:math><tex-math><![CDATA[${J_{C}}=\frac{{a_{C}}}{n-{d_{C}}}$]]></tex-math></alternatives></inline-formula>, where <inline-formula id="j_infor601_ineq_046"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">a</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">C</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${a_{C}}$]]></tex-math></alternatives></inline-formula> is the number of objects classified as 1 by all classifiers from <italic>C</italic>, and <inline-formula id="j_infor601_ineq_047"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">d</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">C</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${d_{C}}$]]></tex-math></alternatives></inline-formula> is the number of objects classified as 0 by all classifiers from <italic>C</italic>.</p>
</list-item>
<list-item id="j_infor601_li_009">
<label>(3)</label>
<p>Determine the 2-group k-adic Jaccard similarity coefficient as, depending whether one set of classifiers can be considered as the reference, 
<disp-formula id="j_infor601_eq_002">
<label>(2)</label><alternatives><mml:math display="block">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">J</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext mathvariant="italic">group</mml:mtext>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">C</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">C</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">J</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">C</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mrow>
<mml:mstyle displaystyle="false">
<mml:mfrac>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">J</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">C</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub>
<mml:mo>+</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">J</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">C</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mo mathvariant="normal">,</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ {J_{\textit{group}}}({C_{1}},{C_{2}})=\frac{{J_{C}}}{\frac{1}{2}({J_{{C_{1}}}}+{J_{{C_{2}}}})},\]]]></tex-math></alternatives>
</disp-formula> 
or 
<disp-formula id="j_infor601_eq_003">
<label>(3)</label><alternatives><mml:math display="block">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd>
<mml:msubsup>
<mml:mrow>
<mml:mi mathvariant="italic">J</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext mathvariant="italic">group</mml:mtext>
</mml:mrow>
<mml:mrow>
<mml:mtext mathvariant="italic">ref</mml:mtext>
</mml:mrow>
</mml:msubsup>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">C</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">C</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">J</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">C</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">J</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">C</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext mathvariant="italic">ref</mml:mtext>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mspace width="1em"/>
<mml:mtext>where</mml:mtext>
<mml:mspace width="2.5pt"/>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">C</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext mathvariant="italic">ref</mml:mtext>
</mml:mrow>
</mml:msub>
<mml:mspace width="2.5pt"/>
<mml:mtext>is a reference set,</mml:mtext>
<mml:mspace width="2.5pt"/>
<mml:mtext mathvariant="italic">ref</mml:mtext>
<mml:mo stretchy="false">∈</mml:mo>
<mml:mo fence="true" stretchy="false">{</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>2</mml:mn>
<mml:mo fence="true" stretchy="false">}</mml:mo>
<mml:mo>.</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ {J_{\textit{group}}^{\textit{ref}}}({C_{1}},{C_{2}})=\frac{{J_{C}}}{{J_{{C_{\textit{ref}}}}}},\hspace{1em}\text{where}\hspace{2.5pt}{C_{\textit{ref}}}\hspace{2.5pt}\text{is a reference set,}\hspace{2.5pt}\textit{ref}\in \{1,2\}.\]]]></tex-math></alternatives>
</disp-formula>
</p>
</list-item>
</list> 
It can be shown that <inline-formula id="j_infor601_ineq_048"><alternatives><mml:math>
<mml:mn>0</mml:mn>
<mml:mo>⩽</mml:mo>
<mml:msubsup>
<mml:mrow>
<mml:mi mathvariant="italic">J</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext mathvariant="italic">group</mml:mtext>
</mml:mrow>
<mml:mrow>
<mml:mtext mathvariant="italic">ref</mml:mtext>
</mml:mrow>
</mml:msubsup>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">C</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">C</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">J</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext mathvariant="italic">group</mml:mtext>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">C</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">C</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>⩽</mml:mo>
<mml:mn>1</mml:mn></mml:math><tex-math><![CDATA[$0\leqslant {J_{\textit{group}}^{\textit{ref}}}({C_{1}},{C_{2}}),{J_{\textit{group}}}({C_{1}},{C_{2}})\leqslant 1$]]></tex-math></alternatives></inline-formula>. The interpretation of the 2-group k-adic Jaccard coefficient depends on the selected approach. In both cases, the coefficient gives the percentage of a within-set consensus agreement on presence that two classifier sets share. When applying Eq. (<xref rid="j_infor601_eq_002">2</xref>), it represents the percentage of the mean within-set similarity preserved when merging the two sets. In the case of Eq. (<xref rid="j_infor601_eq_003">3</xref>), it is the percentage of the reference within-set similarity preserved when merging the two sets. We present the calculation and interpretation of the <inline-formula id="j_infor601_ineq_049"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">J</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext mathvariant="italic">group</mml:mtext>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">C</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">C</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[${J_{\textit{group}}}({C_{1}},{C_{2}})$]]></tex-math></alternatives></inline-formula> coefficient on two simple examples.</p><statement id="j_infor601_stat_005"><label>Example 1.</label>
<p>Assume that we have two sets of binary classifiers: classifier set <italic>A</italic> comprising classifiers <inline-formula id="j_infor601_ineq_050"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">A</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${A_{1}}$]]></tex-math></alternatives></inline-formula> and <inline-formula id="j_infor601_ineq_051"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">A</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${A_{2}}$]]></tex-math></alternatives></inline-formula>, and a classifier set <italic>B</italic> comprising classifiers <inline-formula id="j_infor601_ineq_052"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">B</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${B_{1}}$]]></tex-math></alternatives></inline-formula>, <inline-formula id="j_infor601_ineq_053"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">B</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${B_{2}}$]]></tex-math></alternatives></inline-formula>, and <inline-formula id="j_infor601_ineq_054"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">B</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>3</mml:mn>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${B_{3}}$]]></tex-math></alternatives></inline-formula>. Further assume that a set of 6 objects is classified on a binary scale as shown in Fig. <xref rid="j_infor601_fig_001">1</xref>. In the figure, objects are represented as rows, classifiers as columns, blue cells indicate a classification of 1 (presence), and white cells indicate a classification of 0 (absence).</p>
<p>
<fig id="j_infor601_fig_001">
<label>Fig. 1</label>
<caption>
<p>Example 1A: Binary classifier sets.</p>
</caption>
<graphic xlink:href="infor601_g001.jpg"/>
</fig>
</p>
<p>We quantify within-set similarity in consensus agreement on presence by calculating the k-adic Jaccard coefficient (see equation (<xref rid="j_infor601_eq_001">1</xref>)). We obtain <inline-formula id="j_infor601_ineq_055"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">J</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">A</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo><mml:mstyle displaystyle="false">
<mml:mfrac>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mn>4</mml:mn>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mo>=</mml:mo>
<mml:mn>25</mml:mn>
<mml:mi mathvariant="normal">%</mml:mi></mml:math><tex-math><![CDATA[${J_{A}}=\frac{1}{4}=25\% $]]></tex-math></alternatives></inline-formula> for classifier set A, and <inline-formula id="j_infor601_ineq_056"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">J</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">B</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo><mml:mstyle displaystyle="false">
<mml:mfrac>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mn>5</mml:mn>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mo>=</mml:mo>
<mml:mn>20</mml:mn>
<mml:mi mathvariant="normal">%</mml:mi></mml:math><tex-math><![CDATA[${J_{B}}=\frac{1}{5}=20\% $]]></tex-math></alternatives></inline-formula> for classifier set B. When merging classifier sets A and B, the within-set similarity in consensus agreement drops to <inline-formula id="j_infor601_ineq_057"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">J</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">A</mml:mi>
<mml:mo>∪</mml:mo>
<mml:mi mathvariant="italic">B</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo><mml:mstyle displaystyle="false">
<mml:mfrac>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mn>6</mml:mn>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mo stretchy="false">≈</mml:mo>
<mml:mn>16.67</mml:mn>
<mml:mi mathvariant="normal">%</mml:mi>
<mml:mo>.</mml:mo></mml:math><tex-math><![CDATA[${J_{A\cup B}}=\frac{1}{6}\approx 16.67\% .$]]></tex-math></alternatives></inline-formula> To determine the similarity in consensus agreement on presence between classifier sets <italic>A</italic> and <italic>B</italic>, let’s first consider that there is no reference classifier (see Eq. (<xref rid="j_infor601_eq_002">2</xref>)). We obtain <inline-formula id="j_infor601_ineq_058"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">J</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext mathvariant="italic">group</mml:mtext>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">A</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">B</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo><mml:mstyle displaystyle="false">
<mml:mfrac>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">J</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">A</mml:mi>
<mml:mo>∪</mml:mo>
<mml:mi mathvariant="italic">B</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mrow>
<mml:mstyle displaystyle="false">
<mml:mfrac>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">J</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">A</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>+</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">J</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">B</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mo>=</mml:mo><mml:mstyle displaystyle="false">
<mml:mfrac>
<mml:mrow>
<mml:mstyle displaystyle="false">
<mml:mfrac>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mn>6</mml:mn>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
</mml:mrow>
<mml:mrow>
<mml:mstyle displaystyle="false">
<mml:mfrac>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo><mml:mstyle displaystyle="false">
<mml:mfrac>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mn>4</mml:mn>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mo>+</mml:mo><mml:mstyle displaystyle="false">
<mml:mfrac>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mn>5</mml:mn>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mo stretchy="false">≈</mml:mo>
<mml:mn>74.1</mml:mn>
<mml:mi mathvariant="normal">%</mml:mi>
<mml:mo>.</mml:mo></mml:math><tex-math><![CDATA[${J_{\textit{group}}}(A,B)=\frac{{J_{A\cup B}}}{\frac{1}{2}({J_{A}}+{J_{B}})}=\frac{\frac{1}{6}}{\frac{1}{2}(\frac{1}{4}+\frac{1}{5})}\approx 74.1\% .$]]></tex-math></alternatives></inline-formula> This means that when merging classifier sets <italic>A</italic> and <italic>B</italic>, 74.1% of the mean within-set similarity in consensus agreement of <italic>A</italic> and <italic>B</italic> would be preserved (i.e. <inline-formula id="j_infor601_ineq_059"><alternatives><mml:math>
<mml:mn>25.9</mml:mn>
<mml:mi mathvariant="normal">%</mml:mi></mml:math><tex-math><![CDATA[$25.9\% $]]></tex-math></alternatives></inline-formula> would be lost). Now assume that classifier <italic>A</italic> is a reference. We have <inline-formula id="j_infor601_ineq_060"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">J</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext mathvariant="italic">group</mml:mtext>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">A</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">B</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo><mml:mstyle displaystyle="false">
<mml:mfrac>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">J</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">A</mml:mi>
<mml:mo>∪</mml:mo>
<mml:mi mathvariant="italic">B</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">J</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">A</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mo>=</mml:mo><mml:mstyle displaystyle="false">
<mml:mfrac>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mn>3</mml:mn>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mo stretchy="false">≈</mml:mo>
<mml:mn>66.7</mml:mn>
<mml:mi mathvariant="normal">%</mml:mi></mml:math><tex-math><![CDATA[${J_{\textit{group}}}(A,B)=\frac{{J_{A\cup B}}}{{J_{A}}}=\frac{2}{3}\approx 66.7\% $]]></tex-math></alternatives></inline-formula> (see Eq. (<xref rid="j_infor601_eq_003">3</xref>)). It means that <inline-formula id="j_infor601_ineq_061"><alternatives><mml:math>
<mml:mn>66.7</mml:mn>
<mml:mi mathvariant="normal">%</mml:mi></mml:math><tex-math><![CDATA[$66.7\% $]]></tex-math></alternatives></inline-formula> of the within-set similarity in consensus agreement of classifier set A would be preserved when merging classifier sets <italic>A</italic> and <italic>B</italic>. Similarly, when <italic>B</italic> is the reference group, we have <inline-formula id="j_infor601_ineq_062"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">J</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext mathvariant="italic">group</mml:mtext>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">A</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">B</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo><mml:mstyle displaystyle="false">
<mml:mfrac>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">J</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">A</mml:mi>
<mml:mo>∪</mml:mo>
<mml:mi mathvariant="italic">B</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">J</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">A</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>+</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">J</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">B</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mo>=</mml:mo><mml:mstyle displaystyle="false">
<mml:mfrac>
<mml:mrow>
<mml:mn>5</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mn>6</mml:mn>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mo stretchy="false">≈</mml:mo>
<mml:mn>83.3</mml:mn>
<mml:mi mathvariant="normal">%</mml:mi></mml:math><tex-math><![CDATA[${J_{\textit{group}}}(A,B)=\frac{{J_{A\cup B}}}{{J_{A}}+{J_{B}}}=\frac{5}{6}\approx 83.3\% $]]></tex-math></alternatives></inline-formula>.</p>
<p>Now, consider a slightly different case where two classifiers have both consensus agreement on presence (row 1) and consensus agreement on absence (row 3) for the same sets of objects, as presented in Fig. <xref rid="j_infor601_fig_002">2</xref>.</p>
<p>
<fig id="j_infor601_fig_002">
<label>Fig. 2</label>
<caption>
<p>Example 1B: Binary classifier sets.</p>
</caption>
<graphic xlink:href="infor601_g002.jpg"/>
</fig>
</p>
<p>Here we have <inline-formula id="j_infor601_ineq_063"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">J</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">A</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">J</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">B</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">J</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">A</mml:mi>
<mml:mo>∪</mml:mo>
<mml:mi mathvariant="italic">B</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo><mml:mstyle displaystyle="false">
<mml:mfrac>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mn>5</mml:mn>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mo>=</mml:mo>
<mml:mn>20</mml:mn>
<mml:mi mathvariant="normal">%</mml:mi></mml:math><tex-math><![CDATA[${J_{A}}={J_{B}}={J_{A\cup B}}=\frac{1}{5}=20\% $]]></tex-math></alternatives></inline-formula>. Thus, <inline-formula id="j_infor601_ineq_064"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">J</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext mathvariant="italic">group</mml:mtext>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">A</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">B</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:msubsup>
<mml:mrow>
<mml:mi mathvariant="italic">J</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext mathvariant="italic">group</mml:mtext>
</mml:mrow>
<mml:mrow>
<mml:mtext mathvariant="italic">ref</mml:mtext>
</mml:mrow>
</mml:msubsup>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">A</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">B</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn></mml:math><tex-math><![CDATA[${J_{\textit{group}}}(A,B)={J_{\textit{group}}^{\textit{ref}}}(A,B)=1$]]></tex-math></alternatives></inline-formula>, regardless of the group taken as a reference. Since <inline-formula id="j_infor601_ineq_065"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">J</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext mathvariant="italic">group</mml:mtext>
</mml:mrow>
</mml:msub>
<mml:mo>⩽</mml:mo>
<mml:mn>1</mml:mn></mml:math><tex-math><![CDATA[${J_{\textit{group}}}\leqslant 1$]]></tex-math></alternatives></inline-formula>, this is the highest possible level of similarity in consensus agreement on presence, and thus <inline-formula id="j_infor601_ineq_066"><alternatives><mml:math>
<mml:mn>100</mml:mn>
<mml:mi mathvariant="normal">%</mml:mi></mml:math><tex-math><![CDATA[$100\% $]]></tex-math></alternatives></inline-formula> of the similarity of the consensus agreement on presence would be preserved (<inline-formula id="j_infor601_ineq_067"><alternatives><mml:math>
<mml:mn>0</mml:mn>
<mml:mi mathvariant="normal">%</mml:mi></mml:math><tex-math><![CDATA[$0\% $]]></tex-math></alternatives></inline-formula> would be lost) when merging classifier sets <italic>A</italic> and <italic>B</italic>.</p></statement>
<p>In the application part, none of the classifier set is considered as a reference (Eq. (<xref rid="j_infor601_eq_002">2</xref>)). The properties of that 2-group k-adic Jaccard coefficient and the derivation of the associated confidence interval can be found in Perišić and Vanbelle (<xref ref-type="bibr" rid="j_infor601_ref_017">2024</xref>).</p>
</sec>
<sec id="j_infor601_s_005">
<label>3.2</label>
<title>Quantifying Binary Classifier Algorithms Similarity</title>
<p>In this Section we present the framework for measuring similarity in consensus agreement between two binary classification algorithms. We will consider each binary classification algorithm to be a set of <inline-formula id="j_infor601_ineq_068"><alternatives><mml:math>
<mml:mi mathvariant="italic">k</mml:mi>
<mml:mo>⩾</mml:mo>
<mml:mn>2</mml:mn></mml:math><tex-math><![CDATA[$k\geqslant 2$]]></tex-math></alternatives></inline-formula> binary classifiers, obtained by changing hyperparameter values/subjective inputs. Let <inline-formula id="j_infor601_ineq_069"><alternatives><mml:math>
<mml:mi mathvariant="script">H</mml:mi></mml:math><tex-math><![CDATA[$\mathcal{H}$]]></tex-math></alternatives></inline-formula> be a set of hyperparameter values selected for a certain binary classification algorithm. In general, we are evaluating a set of classifiers <inline-formula id="j_infor601_ineq_070"><alternatives><mml:math>
<mml:mi mathvariant="italic">C</mml:mi>
<mml:mo>=</mml:mo>
<mml:mo fence="true" stretchy="false">{</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">c</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">h</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">h</mml:mi>
<mml:mo stretchy="false">∈</mml:mo>
<mml:mi mathvariant="script">H</mml:mi>
<mml:mo fence="true" stretchy="false">}</mml:mo></mml:math><tex-math><![CDATA[$C=\{{c_{h}},h\in \mathcal{H}\}$]]></tex-math></alternatives></inline-formula>, where <inline-formula id="j_infor601_ineq_071"><alternatives><mml:math>
<mml:mi mathvariant="script">H</mml:mi></mml:math><tex-math><![CDATA[$\mathcal{H}$]]></tex-math></alternatives></inline-formula> is an index set. In this paper, we are evaluating algorithms through a collection of classifiers obtained by applying different hyperparameter values or subjective decisions when adjusting the algorithm. So <inline-formula id="j_infor601_ineq_072"><alternatives><mml:math>
<mml:mi mathvariant="script">H</mml:mi></mml:math><tex-math><![CDATA[$\mathcal{H}$]]></tex-math></alternatives></inline-formula> will be related to the set of hyperparameter values of interest. Although a binary classifier set can be created in an arbitrary way, we consider that there is some reasonable logic behind the creation of such a collection. For instance, a classifier set can be created by evaluating classification rules set by experts from a certain field or institution. We introduce some examples of such collections. Note that we use the term feature to represent traits and variables.</p><statement id="j_infor601_stat_006"><label>Example 2.</label>
<p>Let <inline-formula id="j_infor601_ineq_073"><alternatives><mml:math>
<mml:mi mathvariant="italic">O</mml:mi>
<mml:mo>=</mml:mo>
<mml:mo fence="true" stretchy="false">{</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">o</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo>…</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">o</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo fence="true" stretchy="false">}</mml:mo></mml:math><tex-math><![CDATA[$O=\{{o_{1}},\dots ,{o_{n}}\}$]]></tex-math></alternatives></inline-formula> be a set of objects/items/subjects and <inline-formula id="j_infor601_ineq_074"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo>…</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${X_{1}},\dots ,{X_{p}}$]]></tex-math></alternatives></inline-formula> a set of features. The objects from <italic>O</italic> will be classified on a binary scale by applying a certain algorithm on the basis of the feature values. In other words, we are working with a sample of <italic>n</italic> tuples <inline-formula id="j_infor601_ineq_075"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo>…</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">c</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$({x_{1}}(i),{x_{2}}(i),\dots ,{x_{p}}(i),c(i))$]]></tex-math></alternatives></inline-formula>, <inline-formula id="j_infor601_ineq_076"><alternatives><mml:math>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>2</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo>…</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">n</mml:mi></mml:math><tex-math><![CDATA[$i=1,2,\dots ,n$]]></tex-math></alternatives></inline-formula>, where <inline-formula id="j_infor601_ineq_077"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo>…</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[${x_{1}}(i),{x_{2}}(i),\dots ,{x_{p}}(i)$]]></tex-math></alternatives></inline-formula> represents the observed feature values and <inline-formula id="j_infor601_ineq_078"><alternatives><mml:math>
<mml:mi mathvariant="italic">c</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo stretchy="false">∈</mml:mo>
<mml:mo fence="true" stretchy="false">{</mml:mo>
<mml:mn>0</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo fence="true" stretchy="false">}</mml:mo></mml:math><tex-math><![CDATA[$c(i)\in \{0,1\}$]]></tex-math></alternatives></inline-formula> is the predicted class for the object <inline-formula id="j_infor601_ineq_079"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">o</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${o_{i}}$]]></tex-math></alternatives></inline-formula>. 
<list>
<list-item id="j_infor601_li_010">
<label>(E1)</label>
<p>Consider a classifier set <inline-formula id="j_infor601_ineq_080"><alternatives><mml:math>
<mml:mi mathvariant="italic">C</mml:mi>
<mml:mo>=</mml:mo>
<mml:mo fence="true" stretchy="false">{</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">c</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">c</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">c</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>3</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo fence="true" stretchy="false">}</mml:mo></mml:math><tex-math><![CDATA[$C=\{{c_{1}},{c_{2}},{c_{3}}\}$]]></tex-math></alternatives></inline-formula> of three logistic regression classifiers where <inline-formula id="j_infor601_ineq_081"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">c</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${c_{1}}$]]></tex-math></alternatives></inline-formula> represents a model with all features included, <inline-formula id="j_infor601_ineq_082"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">c</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${c_{2}}$]]></tex-math></alternatives></inline-formula> represents a model with a <inline-formula id="j_infor601_ineq_083"><alternatives><mml:math>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal">&lt;</mml:mo>
<mml:mi mathvariant="italic">k</mml:mi>
<mml:mo mathvariant="normal">&lt;</mml:mo>
<mml:mi mathvariant="italic">p</mml:mi></mml:math><tex-math><![CDATA[$1\lt k\lt p$]]></tex-math></alternatives></inline-formula> features included (for instance, a model built in a stepwise fashion), and <inline-formula id="j_infor601_ineq_084"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">c</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>3</mml:mn>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${c_{3}}$]]></tex-math></alternatives></inline-formula> is a univariate model that includes only one feature with the highest (univariate) prediction performance.</p>
<p>
<fig id="j_infor601_fig_003">
<label>Fig. 3</label>
<caption>
<p>Two classifier sets example. <inline-formula id="j_infor601_ineq_085"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo>×</mml:mo>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${X_{n\times p}}$]]></tex-math></alternatives></inline-formula> presents the data matrix comprising values of <italic>p</italic> features for <italic>n</italic> objects that are classified by applying two algorithms; each algorithm takes into account different hyperparameters, thus comprises a set of classifiers; classification by each algorithm is presented as a matrix where <inline-formula id="j_infor601_ineq_086"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">c</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mi mathvariant="italic">j</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">o</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">l</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[${c_{ij}}({o_{l}})$]]></tex-math></alternatives></inline-formula> presents the predicted classification of an object <inline-formula id="j_infor601_ineq_087"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">o</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">l</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${o_{l}}$]]></tex-math></alternatives></inline-formula> by the classifier <inline-formula id="j_infor601_ineq_088"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">c</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mi mathvariant="italic">j</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${c_{ij}}$]]></tex-math></alternatives></inline-formula>.</p>
</caption>
<graphic xlink:href="infor601_g003.jpg"/>
</fig>
</p>
</list-item>
<list-item id="j_infor601_li_011">
<label>(E2)</label>
<p>Consider a classifier set <inline-formula id="j_infor601_ineq_089"><alternatives><mml:math>
<mml:mi mathvariant="italic">C</mml:mi>
<mml:mo>=</mml:mo>
<mml:mo fence="true" stretchy="false">{</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">c</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mi mathvariant="italic">j</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mspace width="0.1667em"/>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo stretchy="false">∈</mml:mo>
<mml:mi mathvariant="italic">I</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">j</mml:mi>
<mml:mo stretchy="false">∈</mml:mo>
<mml:mi mathvariant="italic">J</mml:mi>
<mml:mo fence="true" stretchy="false">}</mml:mo></mml:math><tex-math><![CDATA[$C=\{{c_{ij}},\hspace{0.1667em}i\in I,j\in J\}$]]></tex-math></alternatives></inline-formula>, where <inline-formula id="j_infor601_ineq_090"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">c</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mi mathvariant="italic">j</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${c_{ij}}$]]></tex-math></alternatives></inline-formula> is a random forest classifier with hyperparameter values <italic>i</italic> (number of trees) and <italic>j</italic> (number of features) that are chosen from a predefined sets of possible values or the values of interest, <italic>I</italic> and <italic>J</italic> respectively.</p>
</list-item>
</list>
</p></statement>
<p>When applied to a dataset of interest, each binary classifier produces a binary output vector, while each algorithm, taking into account different hyperparameter values, produces a set of binary vectors. Figure <xref rid="j_infor601_fig_003">3</xref> presents a visualization of some possible collections of binary classifiers related to two algorithms, applied on a set of objects <inline-formula id="j_infor601_ineq_091"><alternatives><mml:math>
<mml:mo fence="true" stretchy="false">{</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">o</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">o</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo>…</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">o</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo fence="true" stretchy="false">}</mml:mo></mml:math><tex-math><![CDATA[$\{{o_{1}},{o_{2}},\dots ,{o_{n}}\}$]]></tex-math></alternatives></inline-formula>, where each algorithm is represented by a set of classifiers.</p>
<p>Research question (R1) requires assessing the similarity between <italic>k</italic>, <inline-formula id="j_infor601_ineq_092"><alternatives><mml:math>
<mml:mi mathvariant="italic">k</mml:mi>
<mml:mo>⩾</mml:mo>
<mml:mn>2</mml:mn></mml:math><tex-math><![CDATA[$k\geqslant 2$]]></tex-math></alternatives></inline-formula> binary vectors. We propose that for a given dataset of interest and a selected binary classification algorithm, the evaluation of the similarity of classifiers is performed as follows. 
<list>
<list-item id="j_infor601_li_012">
<label><inline-formula id="j_infor601_ineq_093"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mtext mathvariant="italic">Step</mml:mtext>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$({\textit{Step}_{1}})$]]></tex-math></alternatives></inline-formula></label>
<p>Select the set of hyperparemeter values that defines a set of classifiers.</p>
</list-item>
<list-item id="j_infor601_li_013">
<label><inline-formula id="j_infor601_ineq_094"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mtext mathvariant="italic">Step</mml:mtext>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$({\textit{Step}_{2}})$]]></tex-math></alternatives></inline-formula></label>
<p>For each classifier, perform binary classification.</p>
</list-item>
<list-item id="j_infor601_li_014">
<label><inline-formula id="j_infor601_ineq_095"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mtext mathvariant="italic">Step</mml:mtext>
</mml:mrow>
<mml:mrow>
<mml:mn>3</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$({\textit{Step}_{3}})$]]></tex-math></alternatives></inline-formula></label>
<p>Apply the <italic>k</italic>-adic Jaccard coefficient to the set of observed classifications.</p>
</list-item>
</list> 
Research question (R2) requires assessing the similarity between 2 sets or collections of binary vectors. Here, we assume that each binary classification set comprises at least two binary classifiers and propose that for a given dataset and two binary classification algorithms, the similarity between two binary classification algorithms is evaluated as follows. 
<list>
<list-item id="j_infor601_li_015">
<label>(<italic>Step</italic><sub>1</sub>)</label>
<p>For each binary classification algorithm, select the set of hyperparameter values that defines a set of classifiers.</p>
</list-item>
<list-item id="j_infor601_li_016">
<label>(<italic>Step</italic><sub>2</sub>)</label>
<p>For each classifier, perform binary classification. This results with two sets of binary classifications.</p>
</list-item>
<list-item id="j_infor601_li_017">
<label>(<italic>Step</italic><sub>3</sub>)</label>
<p>Apply the 2-group k-adic Jaccard coefficient to the two sets of observed classifications.</p>
</list-item>
</list>
</p>
</sec>
</sec>
<sec id="j_infor601_s_006">
<label>4</label>
<title>Application</title>
<p>This section presents the application of the k-adic Jaccard similarity coefficient and the 2-group k-adic Jaccard similarity coefficient for measuring similarity in consensus agreement for a binary classification algorithm, and between two binary classification algorithms respectively. A part of the calculations was conducted in Petričević (<xref ref-type="bibr" rid="j_infor601_ref_019">2023</xref>). We consider datasets with a binary response variable and different features used as predictors to build a predictive model for classification purposes with the application of different classification algorithms. We involve different hyperparameter values/inputs for each classification algorithm and then evaluate the similarity in consensus agreement. We employ three classification algorithms: logistic regression (LR), random forests (RF) and conditional random forests (CF). The logistic regression models were built with four different sets of predictors. Thus, the first set of binary classifiers comprises four logit-based classifiers. The second set of binary classifiers comprises random forest classifiers built under different hyperparameter values: the number of decision trees in the forest and the number of features considered by each tree when splitting a node. For building prediction models, we used the R package <italic>randomForest</italic> (Liaw and Wiener, <xref ref-type="bibr" rid="j_infor601_ref_013">2002</xref>). The third set of binary classifiers again comprises random forest classifiers, but built under the conditional random forest framework (Strobl <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor601_ref_023">2007</xref>, <xref ref-type="bibr" rid="j_infor601_ref_022">2008</xref>). The resulting forests are unbiased and overcome variable-selection bias which is the major weak spot of the classical approaches. The variable selection bias refers to biasedness in preferring continuous variables and variables with many categories. Again, the models were built under different hyperparameter values: number of decision trees in the forest and the number of features considered by each tree when splitting a node. The models were built using the R package <italic>party</italic> (Strobl <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor601_ref_024">2009</xref>).</p>
<p>We perform the calculations on a balanced dataset and a highly imbalanced dataset by employing two publicly available datasets: <italic>heart disease</italic> (HD) dataset (Janosi <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor601_ref_009">1988</xref>) and <italic>stroke prediction</italic> (SP) dataset (Kaggle, <xref ref-type="bibr" rid="j_infor601_ref_010">2023</xref>). We randomly split both datasets into a train and test dataset where <inline-formula id="j_infor601_ineq_096"><alternatives><mml:math>
<mml:mn>70</mml:mn>
<mml:mi mathvariant="normal">%</mml:mi></mml:math><tex-math><![CDATA[$70\% $]]></tex-math></alternatives></inline-formula> of data are used for training the models. We assess the prediction performance of each classifier on the test dataset by evaluating AUC, accuracy, sensitivity, specificity, precision, and the F1 measure. We need to emphasize that our main goal is not to find the best prediction model, but to construct three collections of predictive models. Thus, some models presented may have lower performance than expected, i.e. by fine-tuning we could find better models. We next describe the results separately for each application.</p>
<sec id="j_infor601_s_007">
<label>4.1</label>
<title>Application I: The HD Dataset</title>
<p>The HD dataset comprises 13 features and 297 instances (after data cleaning). The goal is to build a predictive model, i.e. a binary classifier, for distinguishing presence (classified as 1) and absence (classified as 0) of heart disease. The binary response variable is balanced with <inline-formula id="j_infor601_ineq_097"><alternatives><mml:math>
<mml:mn>46</mml:mn>
<mml:mi mathvariant="normal">%</mml:mi></mml:math><tex-math><![CDATA[$46\% $]]></tex-math></alternatives></inline-formula> of instances being classified as 1 and <inline-formula id="j_infor601_ineq_098"><alternatives><mml:math>
<mml:mn>54</mml:mn>
<mml:mi mathvariant="normal">%</mml:mi></mml:math><tex-math><![CDATA[$54\% $]]></tex-math></alternatives></inline-formula> being classified as 0.</p>
<p>The first set of binary classifiers comprises four logistic regression models (logit based classifiers). The multivariable logit models built for the heart disease dataset were built in a stepwise fashion with a combination of forward and backward selection. The criterion for feature selection was the AIC. We evaluated 4 models: the full model (<italic>lr_heart1</italic>), and three stepwise models with the three lowest AIC values (<italic>lr_heart2</italic>, <italic>lr_heart3</italic>, <italic>lr_heart3</italic>). These models differ in the predictors included as presented in Table <xref rid="j_infor601_tab_001">1</xref>. The disease is classified as present if the estimated probability exceeds the threshold of <inline-formula id="j_infor601_ineq_099"><alternatives><mml:math>
<mml:mi mathvariant="italic">p</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>0.5</mml:mn></mml:math><tex-math><![CDATA[$p=0.5$]]></tex-math></alternatives></inline-formula>.</p>
<table-wrap id="j_infor601_tab_001">
<label>Table 1</label>
<caption>
<p>Predictors in logit models (+ = included, − = excluded) for the HD data.</p>
</caption>
<table>
<thead>
<tr>
<td style="vertical-align: top; text-align: left; border-top: solid thin"/>
<td colspan="4" style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">Logit models</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">Predictor</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin"><italic>lr_heart1</italic></td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin"><italic>lr_heart2</italic></td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin"><italic>lr_heart3</italic></td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin"><italic>lr_heart4</italic></td>
</tr>
</thead>
<tbody>
<tr>
<td style="vertical-align: top; text-align: left"><italic>age</italic></td>
<td style="vertical-align: top; text-align: left">+</td>
<td style="vertical-align: top; text-align: left">−</td>
<td style="vertical-align: top; text-align: left">−</td>
<td style="vertical-align: top; text-align: left">−</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"><italic>sex</italic></td>
<td style="vertical-align: top; text-align: left">+</td>
<td style="vertical-align: top; text-align: left">+</td>
<td style="vertical-align: top; text-align: left">+</td>
<td style="vertical-align: top; text-align: left">+</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"><italic>cp</italic></td>
<td style="vertical-align: top; text-align: left">+</td>
<td style="vertical-align: top; text-align: left">+</td>
<td style="vertical-align: top; text-align: left">+</td>
<td style="vertical-align: top; text-align: left">+</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"><italic>trestbps</italic></td>
<td style="vertical-align: top; text-align: left">+</td>
<td style="vertical-align: top; text-align: left">+</td>
<td style="vertical-align: top; text-align: left">+</td>
<td style="vertical-align: top; text-align: left">+</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"><italic>chol</italic></td>
<td style="vertical-align: top; text-align: left">+</td>
<td style="vertical-align: top; text-align: left">+</td>
<td style="vertical-align: top; text-align: left">+</td>
<td style="vertical-align: top; text-align: left">+</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"><italic>fbs</italic></td>
<td style="vertical-align: top; text-align: left">+</td>
<td style="vertical-align: top; text-align: left">−</td>
<td style="vertical-align: top; text-align: left">−</td>
<td style="vertical-align: top; text-align: left">−</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"><italic>restecg</italic></td>
<td style="vertical-align: top; text-align: left">+</td>
<td style="vertical-align: top; text-align: left">−</td>
<td style="vertical-align: top; text-align: left">−</td>
<td style="vertical-align: top; text-align: left">−</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"><italic>thalach</italic></td>
<td style="vertical-align: top; text-align: left">+</td>
<td style="vertical-align: top; text-align: left">−</td>
<td style="vertical-align: top; text-align: left">+</td>
<td style="vertical-align: top; text-align: left">−</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"><italic>exang</italic></td>
<td style="vertical-align: top; text-align: left">+</td>
<td style="vertical-align: top; text-align: left">+</td>
<td style="vertical-align: top; text-align: left">+</td>
<td style="vertical-align: top; text-align: left">−</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"><italic>oldpeak</italic></td>
<td style="vertical-align: top; text-align: left">+</td>
<td style="vertical-align: top; text-align: left">−</td>
<td style="vertical-align: top; text-align: left">−</td>
<td style="vertical-align: top; text-align: left">−</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"><italic>slope</italic></td>
<td style="vertical-align: top; text-align: left">+</td>
<td style="vertical-align: top; text-align: left">+</td>
<td style="vertical-align: top; text-align: left">+</td>
<td style="vertical-align: top; text-align: left">+</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"><italic>ca</italic></td>
<td style="vertical-align: top; text-align: left">+</td>
<td style="vertical-align: top; text-align: left">+</td>
<td style="vertical-align: top; text-align: left">+</td>
<td style="vertical-align: top; text-align: left">+</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin"><italic>thal</italic></td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">+</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">+</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">+</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">+</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>The second set of binary classifiers comprises nine random forest classifiers built under different hyperparameter values: number of decision trees in the forest <inline-formula id="j_infor601_ineq_100"><alternatives><mml:math>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo stretchy="false">∈</mml:mo>
<mml:mo fence="true" stretchy="false">{</mml:mo>
<mml:mn>50</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>200</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>500</mml:mn>
<mml:mo fence="true" stretchy="false">}</mml:mo></mml:math><tex-math><![CDATA[$n\in \{50,200,500\}$]]></tex-math></alternatives></inline-formula>, and the number of features considered by each tree when splitting a node <inline-formula id="j_infor601_ineq_101"><alternatives><mml:math>
<mml:mi mathvariant="italic">m</mml:mi>
<mml:mo stretchy="false">∈</mml:mo>
<mml:mo fence="true" stretchy="false">{</mml:mo>
<mml:mn>2</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>4</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>10</mml:mn>
<mml:mo fence="true" stretchy="false">}</mml:mo></mml:math><tex-math><![CDATA[$m\in \{2,4,10\}$]]></tex-math></alternatives></inline-formula>.</p>
<p>The third set of binary classifiers comprises nine random forest classifiers built under the conditional random forest framework with different hyperparameter values: number of decision trees in the forest <inline-formula id="j_infor601_ineq_102"><alternatives><mml:math>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo stretchy="false">∈</mml:mo>
<mml:mo fence="true" stretchy="false">{</mml:mo>
<mml:mn>50</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>200</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>500</mml:mn>
<mml:mo fence="true" stretchy="false">}</mml:mo></mml:math><tex-math><![CDATA[$n\in \{50,200,500\}$]]></tex-math></alternatives></inline-formula>, and the number of features considered by each tree when splitting a node <inline-formula id="j_infor601_ineq_103"><alternatives><mml:math>
<mml:mi mathvariant="italic">m</mml:mi>
<mml:mo stretchy="false">∈</mml:mo>
<mml:mo fence="true" stretchy="false">{</mml:mo>
<mml:mn>2</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>4</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>10</mml:mn>
<mml:mo fence="true" stretchy="false">}</mml:mo></mml:math><tex-math><![CDATA[$m\in \{2,4,10\}$]]></tex-math></alternatives></inline-formula>. The performance metrics for the three sets of classifiers are summarized in Table <xref rid="j_infor601_tab_002">2</xref>.</p>
<table-wrap id="j_infor601_tab_002">
<label>Table 2</label>
<caption>
<p>Model valuation for the HD data; cv presents the coefficient of variation for a prediction metric over a set of classifiers.</p>
</caption>
<table>
<thead>
<tr>
<td colspan="3" style="vertical-align: top; text-align: center; border-top: solid thin; border-bottom: solid thin">Model</td>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">Accuracy</td>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">Sensitivity</td>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">Specificity</td>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">Precision</td>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">F1</td>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">AUC</td>
</tr>
</thead>
<tbody>
<tr>
<td rowspan="5" style="vertical-align: top; text-align: left"><italic>LR</italic></td>
<td colspan="2" style="vertical-align: top; text-align: center"><italic>lr_heart1</italic></td>
<td style="vertical-align: top; text-align: left">0.82</td>
<td style="vertical-align: top; text-align: left">0.85</td>
<td style="vertical-align: top; text-align: left">0.79</td>
<td style="vertical-align: top; text-align: left">0.79</td>
<td style="vertical-align: top; text-align: left">0.82</td>
<td style="vertical-align: top; text-align: left">0.82</td>
</tr>
<tr>
<td colspan="2" style="vertical-align: top; text-align: center"><italic>lr_heart2</italic></td>
<td style="vertical-align: top; text-align: left">0.80</td>
<td style="vertical-align: top; text-align: left">0.85</td>
<td style="vertical-align: top; text-align: left">0.74</td>
<td style="vertical-align: top; text-align: left">0.76</td>
<td style="vertical-align: top; text-align: left">0.80</td>
<td style="vertical-align: top; text-align: left">0.80</td>
</tr>
<tr>
<td colspan="2" style="vertical-align: top; text-align: center"><italic>lr_heart3</italic></td>
<td style="vertical-align: top; text-align: left">0.81</td>
<td style="vertical-align: top; text-align: left">0.85</td>
<td style="vertical-align: top; text-align: left">0.77</td>
<td style="vertical-align: top; text-align: left">0.77</td>
<td style="vertical-align: top; text-align: left">0.81</td>
<td style="vertical-align: top; text-align: left">0.81</td>
</tr>
<tr>
<td colspan="2" style="vertical-align: top; text-align: center"><italic>lr_heart4</italic></td>
<td style="vertical-align: top; text-align: left">0.80</td>
<td style="vertical-align: top; text-align: left">0.83</td>
<td style="vertical-align: top; text-align: left">0.77</td>
<td style="vertical-align: top; text-align: left">0.77</td>
<td style="vertical-align: top; text-align: left">0.80</td>
<td style="vertical-align: top; text-align: left">0.80</td>
</tr>
<tr>
<td colspan="2" style="vertical-align: top; text-align: center">cv (%)</td>
<td style="vertical-align: top; text-align: left">1.24</td>
<td style="vertical-align: top; text-align: left">1.28</td>
<td style="vertical-align: top; text-align: left">2.14</td>
<td style="vertical-align: top; text-align: left">1.68</td>
<td style="vertical-align: top; text-align: left">1.15</td>
<td style="vertical-align: top; text-align: left">1.22</td>
</tr>
<tr>
<td rowspan="10" style="vertical-align: top; text-align: left"><italic>CF</italic></td>
<td rowspan="3" style="vertical-align: top; text-align: left"><inline-formula id="j_infor601_ineq_104"><alternatives><mml:math>
<mml:mi mathvariant="italic">m</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>2</mml:mn></mml:math><tex-math><![CDATA[$m=2$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor601_ineq_105"><alternatives><mml:math>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>50</mml:mn></mml:math><tex-math><![CDATA[$n=50$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left">0.82</td>
<td style="vertical-align: top; text-align: left">0.88</td>
<td style="vertical-align: top; text-align: left">0.77</td>
<td style="vertical-align: top; text-align: left">0.78</td>
<td style="vertical-align: top; text-align: left">0.82</td>
<td style="vertical-align: top; text-align: left">0.82</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor601_ineq_106"><alternatives><mml:math>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>200</mml:mn></mml:math><tex-math><![CDATA[$n=200$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left">0.80</td>
<td style="vertical-align: top; text-align: left">0.85</td>
<td style="vertical-align: top; text-align: left">0.74</td>
<td style="vertical-align: top; text-align: left">0.76</td>
<td style="vertical-align: top; text-align: left">0.80</td>
<td style="vertical-align: top; text-align: left">0.80</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor601_ineq_107"><alternatives><mml:math>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>500</mml:mn></mml:math><tex-math><![CDATA[$n=500$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left">0.80</td>
<td style="vertical-align: top; text-align: left">0.85</td>
<td style="vertical-align: top; text-align: left">0.74</td>
<td style="vertical-align: top; text-align: left">0.76</td>
<td style="vertical-align: top; text-align: left">0.80</td>
<td style="vertical-align: top; text-align: left">0.80</td>
</tr>
<tr>
<td rowspan="3" style="vertical-align: top; text-align: left"><inline-formula id="j_infor601_ineq_108"><alternatives><mml:math>
<mml:mi mathvariant="italic">m</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>4</mml:mn></mml:math><tex-math><![CDATA[$m=4$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor601_ineq_109"><alternatives><mml:math>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>50</mml:mn></mml:math><tex-math><![CDATA[$n=50$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left">0.81</td>
<td style="vertical-align: top; text-align: left">0.85</td>
<td style="vertical-align: top; text-align: left">0.77</td>
<td style="vertical-align: top; text-align: left">0.77</td>
<td style="vertical-align: top; text-align: left">0.81</td>
<td style="vertical-align: top; text-align: left">0.81</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor601_ineq_110"><alternatives><mml:math>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>200</mml:mn></mml:math><tex-math><![CDATA[$n=200$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left">0.82</td>
<td style="vertical-align: top; text-align: left">0.85</td>
<td style="vertical-align: top; text-align: left">0.79</td>
<td style="vertical-align: top; text-align: left">0.79</td>
<td style="vertical-align: top; text-align: left">0.82</td>
<td style="vertical-align: top; text-align: left">0.82</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor601_ineq_111"><alternatives><mml:math>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>500</mml:mn></mml:math><tex-math><![CDATA[$n=500$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left">0.77</td>
<td style="vertical-align: top; text-align: left">0.85</td>
<td style="vertical-align: top; text-align: left">0.70</td>
<td style="vertical-align: top; text-align: left">0.72</td>
<td style="vertical-align: top; text-align: left">0.78</td>
<td style="vertical-align: top; text-align: left">0.77</td>
</tr>
<tr>
<td rowspan="3" style="vertical-align: top; text-align: left"><inline-formula id="j_infor601_ineq_112"><alternatives><mml:math>
<mml:mi mathvariant="italic">m</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>10</mml:mn></mml:math><tex-math><![CDATA[$m=10$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor601_ineq_113"><alternatives><mml:math>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>50</mml:mn></mml:math><tex-math><![CDATA[$n=50$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left">0.78</td>
<td style="vertical-align: top; text-align: left">0.85</td>
<td style="vertical-align: top; text-align: left">0.72</td>
<td style="vertical-align: top; text-align: left">0.74</td>
<td style="vertical-align: top; text-align: left">0.79</td>
<td style="vertical-align: top; text-align: left">0.79</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor601_ineq_114"><alternatives><mml:math>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>200</mml:mn></mml:math><tex-math><![CDATA[$n=200$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left">0.78</td>
<td style="vertical-align: top; text-align: left">0.85</td>
<td style="vertical-align: top; text-align: left">0.72</td>
<td style="vertical-align: top; text-align: left">0.74</td>
<td style="vertical-align: top; text-align: left">0.79</td>
<td style="vertical-align: top; text-align: left">0.79</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor601_ineq_115"><alternatives><mml:math>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>500</mml:mn></mml:math><tex-math><![CDATA[$n=500$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left">0.78</td>
<td style="vertical-align: top; text-align: left">0.85</td>
<td style="vertical-align: top; text-align: left">0.72</td>
<td style="vertical-align: top; text-align: left">0.74</td>
<td style="vertical-align: top; text-align: left">0.79</td>
<td style="vertical-align: top; text-align: left">0.79</td>
</tr>
<tr>
<td colspan="2" style="vertical-align: top; text-align: center">cv (%)</td>
<td style="vertical-align: top; text-align: left">2.02</td>
<td style="vertical-align: top; text-align: left">0.92</td>
<td style="vertical-align: top; text-align: left">3.75</td>
<td style="vertical-align: top; text-align: left">2.75</td>
<td style="vertical-align: top; text-align: left">1.67</td>
<td style="vertical-align: top; text-align: left">1.96</td>
</tr>
<tr>
<td rowspan="9" style="vertical-align: top; text-align: left"><italic>RF</italic></td>
<td rowspan="3" style="vertical-align: top; text-align: left"><inline-formula id="j_infor601_ineq_116"><alternatives><mml:math>
<mml:mi mathvariant="italic">m</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>2</mml:mn></mml:math><tex-math><![CDATA[$m=2$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor601_ineq_117"><alternatives><mml:math>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>50</mml:mn></mml:math><tex-math><![CDATA[$n=50$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left">0.81</td>
<td style="vertical-align: top; text-align: left">0.90</td>
<td style="vertical-align: top; text-align: left">0.72</td>
<td style="vertical-align: top; text-align: left">0.75</td>
<td style="vertical-align: top; text-align: left">0.82</td>
<td style="vertical-align: top; text-align: left">0.81</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor601_ineq_118"><alternatives><mml:math>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>200</mml:mn></mml:math><tex-math><![CDATA[$n=200$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left">0.84</td>
<td style="vertical-align: top; text-align: left">0.90</td>
<td style="vertical-align: top; text-align: left">0.79</td>
<td style="vertical-align: top; text-align: left">0.80</td>
<td style="vertical-align: top; text-align: left">0.85</td>
<td style="vertical-align: top; text-align: left">0.85</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor601_ineq_119"><alternatives><mml:math>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>500</mml:mn></mml:math><tex-math><![CDATA[$n=500$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left">0.83</td>
<td style="vertical-align: top; text-align: left">0.88</td>
<td style="vertical-align: top; text-align: left">0.79</td>
<td style="vertical-align: top; text-align: left">0.80</td>
<td style="vertical-align: top; text-align: left">0.8</td>
<td style="vertical-align: top; text-align: left">0.83</td>
</tr>
<tr>
<td rowspan="3" style="vertical-align: top; text-align: left"><inline-formula id="j_infor601_ineq_120"><alternatives><mml:math>
<mml:mi mathvariant="italic">m</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>4</mml:mn></mml:math><tex-math><![CDATA[$m=4$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor601_ineq_121"><alternatives><mml:math>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>50</mml:mn></mml:math><tex-math><![CDATA[$n=50$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left">0.84</td>
<td style="vertical-align: top; text-align: left">0.90</td>
<td style="vertical-align: top; text-align: left">0.79</td>
<td style="vertical-align: top; text-align: left">0.80</td>
<td style="vertical-align: top; text-align: left">0.85</td>
<td style="vertical-align: top; text-align: left">0.85</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor601_ineq_122"><alternatives><mml:math>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>200</mml:mn></mml:math><tex-math><![CDATA[$n=200$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left">0.82</td>
<td style="vertical-align: top; text-align: left">0.90</td>
<td style="vertical-align: top; text-align: left">0.74</td>
<td style="vertical-align: top; text-align: left">0.77</td>
<td style="vertical-align: top; text-align: left">0.83</td>
<td style="vertical-align: top; text-align: left">0.82</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor601_ineq_123"><alternatives><mml:math>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>500</mml:mn></mml:math><tex-math><![CDATA[$n=500$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left">0.82</td>
<td style="vertical-align: top; text-align: left">0.88</td>
<td style="vertical-align: top; text-align: left">0.77</td>
<td style="vertical-align: top; text-align: left">0.78</td>
<td style="vertical-align: top; text-align: left">0.82</td>
<td style="vertical-align: top; text-align: left">0.82</td>
</tr>
<tr>
<td rowspan="3" style="vertical-align: top; text-align: left"><inline-formula id="j_infor601_ineq_124"><alternatives><mml:math>
<mml:mi mathvariant="italic">m</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>10</mml:mn></mml:math><tex-math><![CDATA[$m=10$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor601_ineq_125"><alternatives><mml:math>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>50</mml:mn></mml:math><tex-math><![CDATA[$n=50$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left">0.76</td>
<td style="vertical-align: top; text-align: left">0.80</td>
<td style="vertical-align: top; text-align: left">0.72</td>
<td style="vertical-align: top; text-align: left">0.73</td>
<td style="vertical-align: top; text-align: left">0.76</td>
<td style="vertical-align: top; text-align: left">0.76</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor601_ineq_126"><alternatives><mml:math>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>200</mml:mn></mml:math><tex-math><![CDATA[$n=200$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left">0.77</td>
<td style="vertical-align: top; text-align: left">0.83</td>
<td style="vertical-align: top; text-align: left">0.72</td>
<td style="vertical-align: top; text-align: left">0.73</td>
<td style="vertical-align: top; text-align: left">0.78</td>
<td style="vertical-align: top; text-align: left">0.77</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor601_ineq_127"><alternatives><mml:math>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>500</mml:mn></mml:math><tex-math><![CDATA[$n=500$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left">0.77</td>
<td style="vertical-align: top; text-align: left">0.83</td>
<td style="vertical-align: top; text-align: left">0.72</td>
<td style="vertical-align: top; text-align: left">0.73</td>
<td style="vertical-align: top; text-align: left">0.78</td>
<td style="vertical-align: top; text-align: left">0.77</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin"/>
<td colspan="2" style="vertical-align: top; text-align: center; border-bottom: solid thin">cv (%)</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">3.79</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">4.3</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">4.12</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">3.7</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">3.79</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">3.79</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>For each algorithm, we evaluate the within-set similarity in consensus agreement on presence by calculating the k-adic Jaccard coefficient. The results are presented in Table <xref rid="j_infor601_tab_003">3</xref> with 95% confidence intervals presented in brackets. The LR based classifiers have the highest level of within-set similarity, while the RF based classifiers have the lowest level of within-set similarity. This means that classifications predicted by LR classifiers are less affected by subjective decisions on the number of predictors compared to the sensitivity of RF and CF classifications when changing the hyperparameter values, all with respect to consensus agreement on the presence. Also, we can conclude that the RF based classifiers are more sensitive to changing hyperparameter values than the CF based classifiers, since the within-set similarity in consensus agreement is lower for the RF based classifiers. These results are consistent with the dispersion of the performance measures presented in Table <xref rid="j_infor601_tab_002">2</xref>. The accuracy, specificity, precision, AUC and the F1 measure have the highest dispersion level for the RF algorithm, and the lowest for the LR algorithm.</p>
<table-wrap id="j_infor601_tab_003">
<label>Table 3</label>
<caption>
<p>Within-set similarity for the HD data, confidence intervals presented in brackets.</p>
</caption>
<table>
<thead>
<tr>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">Classifier set</td>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin"><italic>k-adic Jaccard</italic></td>
</tr>
</thead>
<tbody>
<tr>
<td style="vertical-align: top; text-align: left"><italic>LR</italic></td>
<td style="vertical-align: top; text-align: left">0.90 <inline-formula id="j_infor601_ineq_128"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>0.81</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>0.99</mml:mn>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$(0.81,0.99)$]]></tex-math></alternatives></inline-formula></td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"><italic>CF</italic></td>
<td style="vertical-align: top; text-align: left">0.83 <inline-formula id="j_infor601_ineq_129"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>0.72</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>0.95</mml:mn>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$(0.72,0.95)$]]></tex-math></alternatives></inline-formula></td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin"><italic>RF</italic></td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">0.77 <inline-formula id="j_infor601_ineq_130"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>0.64</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>0.89</mml:mn>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$(0.64,0.89)$]]></tex-math></alternatives></inline-formula></td>
</tr>
</tbody>
</table>
</table-wrap>
<p>The between-set similarity in consensus agreement on presence is calculated by applying the 2-group k-adic similarity coefficient. Results are presented in Table <xref rid="j_infor601_tab_004">4</xref> with 95% confidence intervals presented in brackets. The similarity in consensus agreement is similar between pairs of algorithms, with slightly lower similarity between LR and CF. This means that merging classifiers LR and CF would result in less preserved mean within-set similarity in consensus agreement than merging classifiers LR and RF or RF and CF.</p>
<table-wrap id="j_infor601_tab_004">
<label>Table 4</label>
<caption>
<p>The between-set similarity in consensus agreement for the HD data.</p>
</caption>
<table>
<thead>
<tr>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">Classifier set</td>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin"><italic>CF</italic></td>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin"><italic>RF</italic></td>
</tr>
</thead>
<tbody>
<tr>
<td style="vertical-align: top; text-align: left"><italic>LR</italic></td>
<td style="vertical-align: top; text-align: left">0.87 <inline-formula id="j_infor601_ineq_131"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>0.79</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>0.95</mml:mn>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$(0.79,0.95)$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left">0.90 <inline-formula id="j_infor601_ineq_132"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>0.83</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>0.97</mml:mn>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$(0.83,0.97)$]]></tex-math></alternatives></inline-formula></td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin"><italic>CF</italic></td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">1</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">0.90 <inline-formula id="j_infor601_ineq_133"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>0.82</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>0.96</mml:mn>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$(0.82,0.96)$]]></tex-math></alternatives></inline-formula></td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
<sec id="j_infor601_s_008">
<label>4.2</label>
<title>Application II: The SP Dataset</title>
<p>The SP dataset comprises 10 features and 4908 instances (after data cleaning). The predictive models are binary classifiers, classifying instances likely to have a stroke as 1 and those unlikely as 0. The dataset is highly imbalanced where <inline-formula id="j_infor601_ineq_134"><alternatives><mml:math>
<mml:mn>4</mml:mn>
<mml:mi mathvariant="normal">%</mml:mi></mml:math><tex-math><![CDATA[$4\% $]]></tex-math></alternatives></inline-formula> of instances are classified as 1 and <inline-formula id="j_infor601_ineq_135"><alternatives><mml:math>
<mml:mn>96</mml:mn>
<mml:mi mathvariant="normal">%</mml:mi></mml:math><tex-math><![CDATA[$96\% $]]></tex-math></alternatives></inline-formula> are classified as 0. We performed a combination of random undersampling and oversampling for the training data for the SP dataset, which resulted in a balanced training set.</p>
<table-wrap id="j_infor601_tab_005">
<label>Table 5</label>
<caption>
<p>Predictors in logit models (+ = included, − = excluded) for the SP dataset.</p>
</caption>
<table>
<thead>
<tr>
<td style="vertical-align: top; text-align: left; border-top: solid thin"/>
<td colspan="4" style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">Model</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">Predictor</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin"><italic>lr_stroke1</italic></td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin"><italic>lr_stroke2</italic></td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin"><italic>lr_stroke3</italic></td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin"><italic>lr_stroke4</italic></td>
</tr>
</thead>
<tbody>
<tr>
<td style="vertical-align: top; text-align: left"><italic>gender</italic></td>
<td style="vertical-align: top; text-align: left">+</td>
<td style="vertical-align: top; text-align: left">−</td>
<td style="vertical-align: top; text-align: left">+</td>
<td style="vertical-align: top; text-align: left">−</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"><italic>age</italic></td>
<td style="vertical-align: top; text-align: left">+</td>
<td style="vertical-align: top; text-align: left">+</td>
<td style="vertical-align: top; text-align: left">+</td>
<td style="vertical-align: top; text-align: left">+</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"><italic>hypertension</italic></td>
<td style="vertical-align: top; text-align: left">+</td>
<td style="vertical-align: top; text-align: left">+</td>
<td style="vertical-align: top; text-align: left">+</td>
<td style="vertical-align: top; text-align: left">+</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"><italic>heart_disease</italic></td>
<td style="vertical-align: top; text-align: left">+</td>
<td style="vertical-align: top; text-align: left">−</td>
<td style="vertical-align: top; text-align: left">−</td>
<td style="vertical-align: top; text-align: left">−</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"><italic>ever_married</italic></td>
<td style="vertical-align: top; text-align: left">+</td>
<td style="vertical-align: top; text-align: left">−</td>
<td style="vertical-align: top; text-align: left">−</td>
<td style="vertical-align: top; text-align: left">−</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"><italic>work_type</italic></td>
<td style="vertical-align: top; text-align: left">+</td>
<td style="vertical-align: top; text-align: left">+</td>
<td style="vertical-align: top; text-align: left">+</td>
<td style="vertical-align: top; text-align: left">+</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"><italic>Residence_type</italic></td>
<td style="vertical-align: top; text-align: left">+</td>
<td style="vertical-align: top; text-align: left">−</td>
<td style="vertical-align: top; text-align: left">−</td>
<td style="vertical-align: top; text-align: left">−</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"><italic>avg_glucose_level</italic></td>
<td style="vertical-align: top; text-align: left">+</td>
<td style="vertical-align: top; text-align: left">+</td>
<td style="vertical-align: top; text-align: left">+</td>
<td style="vertical-align: top; text-align: left">+</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"><italic>bmi</italic></td>
<td style="vertical-align: top; text-align: left">+</td>
<td style="vertical-align: top; text-align: left">+</td>
<td style="vertical-align: top; text-align: left">+</td>
<td style="vertical-align: top; text-align: left">−</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin"><italic>smoking_status</italic></td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">+</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">+</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">+</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">+</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>We build the prediction models for the SP data following a similar approach to that used for the HD data (see <xref rid="j_infor601_s_007">4.1</xref>). The set of the LR based classifiers comprises the full model (<italic>lr_stroke1</italic>), and three stepwise models: the best performing stepwise model constructed with the combination of forward and backward elimination (<italic>lr_stroke2</italic>), the second best model obtained by the stepwise method with backward elimination (<italic>lr_stroke3</italic>), and the second best model obtained by the stepwise method with forward elimination (<italic>lr_stroke4</italic>). These models differ in predictors (features) included, as presented in Table <xref rid="j_infor601_tab_005">5</xref>. The RF and CF classifiers are also built in the same fashion as for the HD dataset (see Appendix <xref rid="j_infor601_s_007">4.1</xref>). The only difference is in the hyperparameter values related to the number of features considered by each tree when splitting a node <inline-formula id="j_infor601_ineq_136"><alternatives><mml:math>
<mml:mi mathvariant="italic">m</mml:mi>
<mml:mo stretchy="false">∈</mml:mo>
<mml:mo fence="true" stretchy="false">{</mml:mo>
<mml:mn>2</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>4</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>8</mml:mn>
<mml:mo fence="true" stretchy="false">}</mml:mo></mml:math><tex-math><![CDATA[$m\in \{2,4,8\}$]]></tex-math></alternatives></inline-formula>. The performance metrics for evaluated classifiers are summarized in Table <xref rid="j_infor601_tab_006">6</xref>.</p>
<table-wrap id="j_infor601_tab_006">
<label>Table 6</label>
<caption>
<p>Model valuation for the SP data.</p>
</caption>
<table>
<thead>
<tr>
<td colspan="3" style="vertical-align: top; text-align: center; border-top: solid thin; border-bottom: solid thin">Model</td>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">Accuracy</td>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">Sensitivity</td>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">Specificity</td>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">Precision</td>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">F1</td>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">AUC</td>
</tr>
</thead>
<tbody>
<tr>
<td rowspan="5" style="vertical-align: top; text-align: left"><italic>LR</italic></td>
<td colspan="2" style="vertical-align: top; text-align: center"><italic>lr_stroke1</italic></td>
<td style="vertical-align: top; text-align: left">0.74</td>
<td style="vertical-align: top; text-align: left">0.74</td>
<td style="vertical-align: top; text-align: left">0.75</td>
<td style="vertical-align: top; text-align: left">0.99</td>
<td style="vertical-align: top; text-align: left">0.85</td>
<td style="vertical-align: top; text-align: left">0.75</td>
</tr>
<tr>
<td colspan="2" style="vertical-align: top; text-align: center"><italic>lr_stroke2</italic></td>
<td style="vertical-align: top; text-align: left">0.75</td>
<td style="vertical-align: top; text-align: left">0.75</td>
<td style="vertical-align: top; text-align: left">0.77</td>
<td style="vertical-align: top; text-align: left">0.99</td>
<td style="vertical-align: top; text-align: left">0.85</td>
<td style="vertical-align: top; text-align: left">0.76</td>
</tr>
<tr>
<td colspan="2" style="vertical-align: top; text-align: center"><italic>lr_stroke3</italic></td>
<td style="vertical-align: top; text-align: left">0.74</td>
<td style="vertical-align: top; text-align: left">0.74</td>
<td style="vertical-align: top; text-align: left">0.74</td>
<td style="vertical-align: top; text-align: left">0.99</td>
<td style="vertical-align: top; text-align: left">0.85</td>
<td style="vertical-align: top; text-align: left">0.74</td>
</tr>
<tr>
<td colspan="2" style="vertical-align: top; text-align: center"><italic>lr_stroke4</italic></td>
<td style="vertical-align: top; text-align: left">0.75</td>
<td style="vertical-align: top; text-align: left">0.75</td>
<td style="vertical-align: top; text-align: left">0.75</td>
<td style="vertical-align: top; text-align: left">0.99</td>
<td style="vertical-align: top; text-align: left">0.85</td>
<td style="vertical-align: top; text-align: left">0.75</td>
</tr>
<tr>
<td colspan="2" style="vertical-align: top; text-align: center">cv (%)</td>
<td style="vertical-align: top; text-align: left">0.37</td>
<td style="vertical-align: top; text-align: left">0.34</td>
<td style="vertical-align: top; text-align: left">1.65</td>
<td style="vertical-align: top; text-align: left">0.07</td>
<td style="vertical-align: top; text-align: left">0.21</td>
<td style="vertical-align: top; text-align: left">0.96</td>
</tr>
<tr>
<td rowspan="10" style="vertical-align: top; text-align: left"><italic>CF</italic></td>
<td rowspan="3" style="vertical-align: top; text-align: left">90<inline-formula id="j_infor601_ineq_137"><alternatives><mml:math>
<mml:mi mathvariant="italic">m</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>2</mml:mn></mml:math><tex-math><![CDATA[$m=2$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor601_ineq_138"><alternatives><mml:math>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>50</mml:mn></mml:math><tex-math><![CDATA[$n=50$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left">0.79</td>
<td style="vertical-align: top; text-align: left">0.80</td>
<td style="vertical-align: top; text-align: left">0.65</td>
<td style="vertical-align: top; text-align: left">0.98</td>
<td style="vertical-align: top; text-align: left">0.88</td>
<td style="vertical-align: top; text-align: left">0.72</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor601_ineq_139"><alternatives><mml:math>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>200</mml:mn></mml:math><tex-math><![CDATA[$n=200$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left">0.79</td>
<td style="vertical-align: top; text-align: left">0.80</td>
<td style="vertical-align: top; text-align: left">0.68</td>
<td style="vertical-align: top; text-align: left">0.98</td>
<td style="vertical-align: top; text-align: left">0.88</td>
<td style="vertical-align: top; text-align: left">0.74</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor601_ineq_140"><alternatives><mml:math>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>500</mml:mn></mml:math><tex-math><![CDATA[$n=500$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left">0.79</td>
<td style="vertical-align: top; text-align: left">0.80</td>
<td style="vertical-align: top; text-align: left">0.68</td>
<td style="vertical-align: top; text-align: left">0.98</td>
<td style="vertical-align: top; text-align: left">0.88</td>
<td style="vertical-align: top; text-align: left">0.74</td>
</tr>
<tr>
<td rowspan="3" style="vertical-align: top; text-align: left"><inline-formula id="j_infor601_ineq_141"><alternatives><mml:math>
<mml:mi mathvariant="italic">m</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>4</mml:mn></mml:math><tex-math><![CDATA[$m=4$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor601_ineq_142"><alternatives><mml:math>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>50</mml:mn></mml:math><tex-math><![CDATA[$n=50$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left">0.82</td>
<td style="vertical-align: top; text-align: left">0.83</td>
<td style="vertical-align: top; text-align: left">0.63</td>
<td style="vertical-align: top; text-align: left">0.98</td>
<td style="vertical-align: top; text-align: left">0.90</td>
<td style="vertical-align: top; text-align: left">0.73</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor601_ineq_143"><alternatives><mml:math>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>200</mml:mn></mml:math><tex-math><![CDATA[$n=200$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left">0.83</td>
<td style="vertical-align: top; text-align: left">0.83</td>
<td style="vertical-align: top; text-align: left">0.67</td>
<td style="vertical-align: top; text-align: left">0.98</td>
<td style="vertical-align: top; text-align: left">0.90</td>
<td style="vertical-align: top; text-align: left">0.75</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor601_ineq_144"><alternatives><mml:math>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>500</mml:mn></mml:math><tex-math><![CDATA[$n=500$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left">0.82</td>
<td style="vertical-align: top; text-align: left">0.83</td>
<td style="vertical-align: top; text-align: left">0.67</td>
<td style="vertical-align: top; text-align: left">0.98</td>
<td style="vertical-align: top; text-align: left">0.90</td>
<td style="vertical-align: top; text-align: left">0.75</td>
</tr>
<tr>
<td rowspan="3" style="vertical-align: top; text-align: left"><inline-formula id="j_infor601_ineq_145"><alternatives><mml:math>
<mml:mi mathvariant="italic">m</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>8</mml:mn></mml:math><tex-math><![CDATA[$m=8$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor601_ineq_146"><alternatives><mml:math>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>50</mml:mn></mml:math><tex-math><![CDATA[$n=50$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left">0.83</td>
<td style="vertical-align: top; text-align: left">0.84</td>
<td style="vertical-align: top; text-align: left">0.58</td>
<td style="vertical-align: top; text-align: left">0.98</td>
<td style="vertical-align: top; text-align: left">0.91</td>
<td style="vertical-align: top; text-align: left">0.71</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor601_ineq_147"><alternatives><mml:math>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>200</mml:mn></mml:math><tex-math><![CDATA[$n=200$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left">0.83</td>
<td style="vertical-align: top; text-align: left">0.84</td>
<td style="vertical-align: top; text-align: left">0.60</td>
<td style="vertical-align: top; text-align: left">0.98</td>
<td style="vertical-align: top; text-align: left">0.90</td>
<td style="vertical-align: top; text-align: left">0.72</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor601_ineq_148"><alternatives><mml:math>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>500</mml:mn></mml:math><tex-math><![CDATA[$n=500$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left">0.83</td>
<td style="vertical-align: top; text-align: left">0.84</td>
<td style="vertical-align: top; text-align: left">0.58</td>
<td style="vertical-align: top; text-align: left">0.98</td>
<td style="vertical-align: top; text-align: left">0.90</td>
<td style="vertical-align: top; text-align: left">0.71</td>
</tr>
<tr>
<td colspan="2" style="vertical-align: top; text-align: center">cv (%)</td>
<td style="vertical-align: top; text-align: left">1.93</td>
<td style="vertical-align: top; text-align: left">2.11</td>
<td style="vertical-align: top; text-align: left">6.36</td>
<td style="vertical-align: top; text-align: left">0.17</td>
<td style="vertical-align: top; text-align: left">1.11</td>
<td style="vertical-align: top; text-align: left">1.91</td>
</tr>
<tr>
<td rowspan="10" style="vertical-align: top; text-align: left; border-bottom: solid thin"><italic>RF</italic></td>
<td rowspan="3" style="vertical-align: top; text-align: left"><inline-formula id="j_infor601_ineq_149"><alternatives><mml:math>
<mml:mi mathvariant="italic">m</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>2</mml:mn></mml:math><tex-math><![CDATA[$m=2$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor601_ineq_150"><alternatives><mml:math>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>50</mml:mn></mml:math><tex-math><![CDATA[$n=50$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left">0.86</td>
<td style="vertical-align: top; text-align: left">0.88</td>
<td style="vertical-align: top; text-align: left">0.40</td>
<td style="vertical-align: top; text-align: left">0.97</td>
<td style="vertical-align: top; text-align: left">0.92</td>
<td style="vertical-align: top; text-align: left">0.64</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor601_ineq_151"><alternatives><mml:math>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>200</mml:mn></mml:math><tex-math><![CDATA[$n=200$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left">0.86</td>
<td style="vertical-align: top; text-align: left">0.88</td>
<td style="vertical-align: top; text-align: left">0.44</td>
<td style="vertical-align: top; text-align: left">0.98</td>
<td style="vertical-align: top; text-align: left">0.92</td>
<td style="vertical-align: top; text-align: left">0.66</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor601_ineq_152"><alternatives><mml:math>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>500</mml:mn></mml:math><tex-math><![CDATA[$n=500$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left">0.86</td>
<td style="vertical-align: top; text-align: left">0.88</td>
<td style="vertical-align: top; text-align: left">0.46</td>
<td style="vertical-align: top; text-align: left">0.98</td>
<td style="vertical-align: top; text-align: left">0.92</td>
<td style="vertical-align: top; text-align: left">0.67</td>
</tr>
<tr>
<td rowspan="3" style="vertical-align: top; text-align: left"><inline-formula id="j_infor601_ineq_153"><alternatives><mml:math>
<mml:mi mathvariant="italic">m</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>4</mml:mn></mml:math><tex-math><![CDATA[$m=4$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor601_ineq_154"><alternatives><mml:math>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>50</mml:mn></mml:math><tex-math><![CDATA[$n=50$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left">0.92</td>
<td style="vertical-align: top; text-align: left">0.95</td>
<td style="vertical-align: top; text-align: left">0.23</td>
<td style="vertical-align: top; text-align: left">0.97</td>
<td style="vertical-align: top; text-align: left">0.96</td>
<td style="vertical-align: top; text-align: left">0.59</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor601_ineq_155"><alternatives><mml:math>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>200</mml:mn></mml:math><tex-math><![CDATA[$n=200$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left">0.92</td>
<td style="vertical-align: top; text-align: left">0.950</td>
<td style="vertical-align: top; text-align: left">0.26</td>
<td style="vertical-align: top; text-align: left">0.97</td>
<td style="vertical-align: top; text-align: left">0.96</td>
<td style="vertical-align: top; text-align: left">0.61</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor601_ineq_156"><alternatives><mml:math>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>500</mml:mn></mml:math><tex-math><![CDATA[$n=500$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left">0.92</td>
<td style="vertical-align: top; text-align: left">0.95</td>
<td style="vertical-align: top; text-align: left">0.25</td>
<td style="vertical-align: top; text-align: left">0.97</td>
<td style="vertical-align: top; text-align: left">0.96</td>
<td style="vertical-align: top; text-align: left">0.60</td>
</tr>
<tr>
<td rowspan="3" style="vertical-align: top; text-align: left"><inline-formula id="j_infor601_ineq_157"><alternatives><mml:math>
<mml:mi mathvariant="italic">m</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>8</mml:mn></mml:math><tex-math><![CDATA[$m=8$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor601_ineq_158"><alternatives><mml:math>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>50</mml:mn></mml:math><tex-math><![CDATA[$n=50$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left">0.91</td>
<td style="vertical-align: top; text-align: left">0.94</td>
<td style="vertical-align: top; text-align: left">0.19</td>
<td style="vertical-align: top; text-align: left">0.97</td>
<td style="vertical-align: top; text-align: left">0.95</td>
<td style="vertical-align: top; text-align: left">0.57</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor601_ineq_159"><alternatives><mml:math>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>200</mml:mn></mml:math><tex-math><![CDATA[$n=200$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left">0.91</td>
<td style="vertical-align: top; text-align: left">0.94</td>
<td style="vertical-align: top; text-align: left">0.23</td>
<td style="vertical-align: top; text-align: left">0.97</td>
<td style="vertical-align: top; text-align: left">0.96</td>
<td style="vertical-align: top; text-align: left">0.58</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor601_ineq_160"><alternatives><mml:math>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>500</mml:mn></mml:math><tex-math><![CDATA[$n=500$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left">0.91</td>
<td style="vertical-align: top; text-align: left">0.94</td>
<td style="vertical-align: top; text-align: left">0.23</td>
<td style="vertical-align: top; text-align: left">0.97</td>
<td style="vertical-align: top; text-align: left">0.95</td>
<td style="vertical-align: top; text-align: left">0.58</td>
</tr>
<tr>
<td colspan="2" style="vertical-align: top; text-align: center; border-bottom: solid thin">cv (%)</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">3.06</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">3.52</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">30.63</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">0.33</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">1.52</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">5.43</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>We evaluate the within-set similarity in consensus agreement on presence by calculating the k-adic Jaccard coefficient. The results are presented in Table <xref rid="j_infor601_tab_007">7</xref>. As in the case of the HD data, the LR based classifiers have the highest level of within-set similarity, while the RF based classifiers have the lowest level of within-set similarity. Again, this means that classifications predicted by LR based classifiers are less affected by subjective decisions on the number of predictors compared to the sensitivity of RF and CF classifications when changing the hyperparameter values, all with respect to consensus agreement on the presence. Also, as in the case of the HD dataset, we can conclude that classifications predicted by RF based classifiers are more sensitive to changing hyperparameter values than classifications predicted by the CF based classifiers, since the within-set similarity in consensus agreement is lower for the RF based classifiers. Compared to the HD dataset case, in the case of the SP data, the RF and CF algorithms exhibit notably higher sensitivity to hyperparameter values (due to the lower values of the similarity coefficient), all in terms of the within-set similarity in consensus agreement. These results are consistent with the dispersion of the performance measures presented in Table <xref rid="j_infor601_tab_006">6</xref>. All performance measures have the highest dispersion level for the RF algorithm, and the lowest for the LR algorithm.</p>
<table-wrap id="j_infor601_tab_007">
<label>Table 7</label>
<caption>
<p>Within-set similarity for the SP data.</p>
</caption>
<table>
<thead>
<tr>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">Classifier set</td>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin"><italic>k-adic Jaccard</italic></td>
</tr>
</thead>
<tbody>
<tr>
<td style="vertical-align: top; text-align: left"><italic>LR</italic></td>
<td style="vertical-align: top; text-align: left">0.92 <inline-formula id="j_infor601_ineq_161"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>0.90</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>0.95</mml:mn>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$(0.90,0.95)$]]></tex-math></alternatives></inline-formula></td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"><italic>CF</italic></td>
<td style="vertical-align: top; text-align: left">0.63 <inline-formula id="j_infor601_ineq_162"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>0.58</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>0.68</mml:mn>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$(0.58,0.68)$]]></tex-math></alternatives></inline-formula></td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin"><italic>RF</italic></td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">0.24 <inline-formula id="j_infor601_ineq_163"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>0.18</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>0.29</mml:mn>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$(0.18,0.29)$]]></tex-math></alternatives></inline-formula></td>
</tr>
</tbody>
</table>
</table-wrap>
<p>The between-set similarity in consensus agreement on presence is calculated by applying the 2-group k-adic similarity coefficient, and the results are presented in Table <xref rid="j_infor601_tab_008">8</xref> with 95% confidence intervals in brackets. LR-based and CF-based algorithms have the highest level of similarity in consensus agreement on presence, while the LR-based algorithm and the RF-based algorithm have the lowest level of similarity in consensus agreement on presence.</p>
<table-wrap id="j_infor601_tab_008">
<label>Table 8</label>
<caption>
<p>The between-set similarity in consensus agreement for the SP data.</p>
</caption>
<table>
<thead>
<tr>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">Classifier set</td>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin"><italic>CF</italic></td>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin"><italic>RF</italic></td>
</tr>
</thead>
<tbody>
<tr>
<td style="vertical-align: top; text-align: left"><italic>LR</italic></td>
<td style="vertical-align: top; text-align: left">0.61 <inline-formula id="j_infor601_ineq_164"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>0.57</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>0.65</mml:mn>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$(0.57,0.65)$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left">0.22 <inline-formula id="j_infor601_ineq_165"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>0.18</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>0.27</mml:mn>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$(0.18,0.27)$]]></tex-math></alternatives></inline-formula></td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin"><italic>CF</italic></td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">1</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">0.36 <inline-formula id="j_infor601_ineq_166"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>0.29</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>0.41</mml:mn>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$(0.29,0.41)$]]></tex-math></alternatives></inline-formula></td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
</sec>
<sec id="j_infor601_s_009">
<label>5</label>
<title>Conclusion and Discussion</title>
<p>Binary classification is a fundamental task in machine learning, with broad applications across various scientific and practical fields. A vast and growing number of developed binary classification algorithms supports the importance and encourages the application of binary classification methods. The extensive collection of developed algorithms challenges researchers to develop methods to find the best-suiting algorithm for a particular problem and the dataset of interest. This typically involves comparing the predictive performance of binary classification models.</p>
<p>Building an effective prediction model typically requires hyperparameter tuning, meaning that when evaluating a single binary classification algorithm with different hyperparameter values, we are effectively assessing a collection of classifiers.</p>
<p>This research makes a significant contribution to the comparison of binary classification algorithms. Instead of focusing on prediction performances, it compares the resulting classifications. The core of our approach lies in considering the variability introduced by different hyperparameter values for each algorithm when performing such comparisons. This means that, rather than selecting a single representative classifier for each binary classification algorithm (typically the one with the best prediction performance), we consider a set of classifiers as being a representative set of a binary classification algorithm. This approach accounts for the variability of that set.</p>
<p>The comparison of two binary classification algorithms is performed by applying the 2-group k-adic similarity coefficient. This approach is based on the k-adic similarity approach, or the De Moivre definition of agreement. The variability within each classifier set is quantified by calculating the k-adic similarity coefficients. Specifically, we apply the k-adic similarity coefficient to assess the within-set similarity, which quantifies the similarity of <italic>k</italic> classifiers related to a particular classification algorithm. Furthermore, when classifiers are built under a certain classification algorithm by varying hyperparameter values and subjective inputs, the k-adic similarity coefficient can be used to assess the sensitivity of an algorithm to changes in these values.</p>
<p>The 2-group k-adic Jaccard coefficient and the k-adic Jaccard coefficient have a broad potential of application which has yet to be explored. For instance, it could be of interest to evaluate the within-set similarity of a classifier set of a classification across different hyperparameter spaces. This can be performed by calculating the k-adic Jaccard coefficient for different <inline-formula id="j_infor601_ineq_167"><alternatives><mml:math>
<mml:mi mathvariant="script">H</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo></mml:math><tex-math><![CDATA[$\mathcal{H},$]]></tex-math></alternatives></inline-formula> and could be used as a measure of robustness. In future research, we plan to relax the strict consensus-based approach of the k-adic similarity coefficient. While this method evaluates the agreement between classifiers, we plan to develop similarity coefficients that can measure partial agreement between two sets of classifiers.</p>
<p>The proposed methodology applies the asymmetric similarity coefficients. However, when the goal of the analysis is to take into account both agreement on presence and agreement on absence, we involve the symmetric coefficients, like the simple matching coefficient. A version of this coefficient that can be applied for quantifying binary classifier algorithms similarity can be found in Perišić and Vanbelle (<xref ref-type="bibr" rid="j_infor601_ref_017">2024</xref>).</p>
<p>The presented methodology has several limitations and raises important open questions. The first one relates to the selection of hyperparameter values. The hyperparameter space is multidimensional, with each dimension representing a specific hyperparameter. The same complexity applies for the space related to subjective input values. Key questions to address include:</p>
<list>
<list-item id="j_infor601_li_018">
<label>•</label>
<p>How many distinct hyperparameters should be chosen when forming a set of binary classifiers for some classification algorithm?</p>
</list-item>
<list-item id="j_infor601_li_019">
<label>•</label>
<p>What criteria should guide the selection of hyperparameter values?</p>
</list-item>
<list-item id="j_infor601_li_020">
<label>•</label>
<p>Is it justified to compare the sensitivity to hyperparameters when different hyperparameters or subjective inputs are considered for each algorithm?</p>
</list-item>
</list>
<p>For example, in the application described in Section 4, logistic regression was found to be less sensitive to variations in hyperparameters and inputs, while random forest was more sensitive. Logistic regression classifiers were created by selecting different predictors during model building, whereas random forests were formed by altering parameters such as the number of trees in the forest and the number of features considered at each split. This suggests that it may not always be appropriate to consider the same hyperparameter dimensions across algorithms. Thus, comparing algorithm sensitivity to changing hyperparameter values will sometimes imply evaluating different hyperparameter dimensions for each classification algorithm. However, we should be aware that results on algorithm sensitivity could be affected by selected hyperparameter dimensions.</p>
<p>Another limitation arises when comparing classifier sets of different sizes. If we assess the within-set similarity of a collection of classifiers derived from various hyperparameter settings for a given algorithm, increasing the number of classifiers in the set (i.e. evaluating a broader range of hyperparameter values) could increase the heterogeneity within the set. This is evident in the examples provided in Appendix <xref rid="j_infor601_app_001">A</xref>. Consequently, the selection of both the number of classifiers and the specific hyperparameter values should be approached with care, particularly if no predefined rule dictates which hyperparameters are of interest.</p>
</sec>
</body>
<back>
<app-group>
<app id="j_infor601_app_001"><label>A</label>
<title>Appendix: On the Number of Evaluated Classifiers</title>
<p>In the application part presented in Section <xref rid="j_infor601_s_006">4</xref>, the number of classifiers related to LR algorithm (i.e. 4) is much smaller than the number of classifiers related to RF and CF (i.e. 9). This can affect the value of the similarity coefficient. We analyse this impact by taking samples of 4 classifiers from the CF and RF based classifier sets separately, and evaluating the within- and the between-set similarity.</p>
<sec id="j_infor601_s_010">
<label>A.1</label>
<title>HD Dataset</title>
<p>The mean value of the 126 k-adic Jaccard coefficients obtained within sets of 4 classifiers is presented in Table <xref rid="j_infor601_tab_009">9</xref>. These means are slightly higher than when considering the set of 9 classifiers.</p>
<table-wrap id="j_infor601_tab_009">
<label>Table 9</label>
<caption>
<p>Within-set similarity for the HD data, mean values of the 4-classifier subsets, the range <inline-formula id="j_infor601_ineq_168"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo movablelimits="false">min</mml:mo>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo movablelimits="false">max</mml:mo>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$({x_{\min }},{x_{\max }})$]]></tex-math></alternatives></inline-formula> presented in brackets.</p>
</caption>
<table>
<thead>
<tr>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">Classifier set</td>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin"><italic>k-adic Jaccard</italic></td>
</tr>
</thead>
<tbody>
<tr>
<td style="vertical-align: top; text-align: left"><italic>CF</italic></td>
<td style="vertical-align: top; text-align: left">0.85 <inline-formula id="j_infor601_ineq_169"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>0.83</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>0.93</mml:mn>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$(0.83,0.93)$]]></tex-math></alternatives></inline-formula></td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin"><italic>RF</italic></td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">0.81 <inline-formula id="j_infor601_ineq_170"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>0.77</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>0.90</mml:mn>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$(0.77,0.90)$]]></tex-math></alternatives></inline-formula></td>
</tr>
</tbody>
</table>
</table-wrap>
<p>We further evaluate the between-set similarities when taking subsets of 4 classifiers from the RF and CF classifier sets separately (i.e. 126 sets for each algorithm) and calculate the mean of the 2-group k-adic Jaccard similarity coefficients (see Table <xref rid="j_infor601_tab_010">10</xref>). The mean coefficient values slightly differ from the case when the 9 classifiers were considered at the same time.</p>
<table-wrap id="j_infor601_tab_010">
<label>Table 10</label>
<caption>
<p>The between-set similarity in consensus agreement for the HD data-mean values of the 4-classifier subsets, the range <inline-formula id="j_infor601_ineq_171"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo movablelimits="false">min</mml:mo>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo movablelimits="false">max</mml:mo>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$({x_{\min }},{x_{\max }})$]]></tex-math></alternatives></inline-formula> presented in brackets.</p>
</caption>
<table>
<thead>
<tr>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">Classifier set</td>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin"><italic>CF</italic></td>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin"><italic>RF</italic></td>
</tr>
</thead>
<tbody>
<tr>
<td style="vertical-align: top; text-align: left"><italic>LR</italic></td>
<td style="vertical-align: top; text-align: left">0.90 <inline-formula id="j_infor601_ineq_172"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>0.86</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>0.96</mml:mn>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$(0.86,0.96)$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left">0.90 <inline-formula id="j_infor601_ineq_173"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>0.87</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>0.93</mml:mn>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$(0.87,0.93)$]]></tex-math></alternatives></inline-formula></td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin"><italic>CF</italic></td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">1</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">0.89 <inline-formula id="j_infor601_ineq_174"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>0.84</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>0.95</mml:mn>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$(0.84,0.95)$]]></tex-math></alternatives></inline-formula></td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
<sec id="j_infor601_s_011">
<label>A.2</label>
<title>SP Dataset</title>
<p>For the SP dataset, when considering sets of 4 classifiers, the mean value of the 126 k-adic Jaccard coefficients (see Table <xref rid="j_infor601_tab_011">11</xref>) is higher than when considering the set of 9 classifiers. Here also, we evaluate the between-set similarity when taking subsets of 4 classifiers from the RF and CF classifier sets (126 sets for each algorithm) and calculate the mean of the 2-group k-adic Jaccard similarity coefficients (see Table <xref rid="j_infor601_tab_012">12</xref>). All mean coefficient values are higher than in the case with 9 classifiers. As in the case of 9 classifiers, in the case of evaluating the mean values of the 4-classifier subsets, RF has the lowest within-set similarity. Also, the lowest between-set similarity is again related to LR and RF, while the highest between-set similarity is related to LR and CF.</p>
<table-wrap id="j_infor601_tab_011">
<label>Table 11</label>
<caption>
<p>Within-set similarity for the SP data, mean values of the 4-classifier subsets, the range <inline-formula id="j_infor601_ineq_175"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo movablelimits="false">min</mml:mo>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo movablelimits="false">max</mml:mo>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$({x_{\min }},{x_{\max }})$]]></tex-math></alternatives></inline-formula> presented in brackets.</p>
</caption>
<table>
<thead>
<tr>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">Classifier set</td>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin"><italic>k-adic Jaccard</italic></td>
</tr>
</thead>
<tbody>
<tr>
<td style="vertical-align: top; text-align: left"><italic>CF</italic></td>
<td style="vertical-align: top; text-align: left">0.71 <inline-formula id="j_infor601_ineq_176"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>0.65</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>0.84</mml:mn>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$(0.65,0.84)$]]></tex-math></alternatives></inline-formula></td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin"><italic>RF</italic></td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">0.34 <inline-formula id="j_infor601_ineq_177"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>0.25</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>0.61</mml:mn>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$(0.25,0.61)$]]></tex-math></alternatives></inline-formula></td>
</tr>
</tbody>
</table>
</table-wrap>
<table-wrap id="j_infor601_tab_012">
<label>Table 12</label>
<caption>
<p>The between-set similarity in consensus agreement for the SP data-mean values of the 4-classifier subsets, the range <inline-formula id="j_infor601_ineq_178"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo movablelimits="false">min</mml:mo>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo movablelimits="false">max</mml:mo>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$({x_{\min }},{x_{\max }})$]]></tex-math></alternatives></inline-formula> presented in brackets.</p>
</caption>
<table>
<thead>
<tr>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">Classifier set</td>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin"><italic>CF</italic></td>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin"><italic>RF</italic></td>
</tr>
</thead>
<tbody>
<tr>
<td style="vertical-align: top; text-align: left"><italic>LR</italic></td>
<td style="vertical-align: top; text-align: left">0.81 <inline-formula id="j_infor601_ineq_179"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>0.69</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>0.89</mml:mn>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$(0.69,0.89)$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left">0.24 <inline-formula id="j_infor601_ineq_180"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>0.19</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>0.30</mml:mn>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$(0.19,0.30)$]]></tex-math></alternatives></inline-formula></td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin"><italic>CF</italic></td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">1</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">0.37 <inline-formula id="j_infor601_ineq_181"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>0.26</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>0.46</mml:mn>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$(0.26,0.46)$]]></tex-math></alternatives></inline-formula></td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
</app></app-group>
<ref-list id="j_infor601_reflist_001">
<title>References</title>
<ref id="j_infor601_ref_001">
<mixed-citation publication-type="journal"><string-name><surname>Benavoli</surname>, <given-names>A.</given-names></string-name>, <string-name><surname>Corani</surname>, <given-names>G.</given-names></string-name>, <string-name><surname>Mangili</surname>, <given-names>F.</given-names></string-name> (<year>2016</year>). <article-title>Should we really use post-hoc tests based on mean-ranks?</article-title> <source>The Journal of Machine Learning Research</source>, <volume>17</volume>(<issue>1</issue>), <fpage>152</fpage>–<lpage>161</lpage>.</mixed-citation>
</ref>
<ref id="j_infor601_ref_002">
<mixed-citation publication-type="journal"><string-name><surname>Bischl</surname>, <given-names>B.</given-names></string-name>, <string-name><surname>Binder</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Lang</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Pielok</surname>, <given-names>T.</given-names></string-name>, <string-name><surname>Richter</surname>, <given-names>J.</given-names></string-name>, <string-name><surname>Coors</surname>, <given-names>S.</given-names></string-name>, <string-name><surname>Thomas</surname>, <given-names>J.</given-names></string-name>, <string-name><surname>Ullmann</surname>, <given-names>T.</given-names></string-name>, <string-name><surname>Becker</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Boulesteix</surname>, <given-names>A.-L.</given-names></string-name>, <string-name><surname>Deng</surname>, <given-names>D.</given-names></string-name>, <string-name><surname>Lindauer</surname>, <given-names>M.</given-names></string-name> (<year>2023</year>). <article-title>Hyperparameter optimization: foundations, algorithms, best practices, and open challenges</article-title>. <source>WIREs Data Mining and Knowledge Discovery</source>, <volume>13</volume>(<issue>2</issue>), <elocation-id>e1484</elocation-id>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1002/widm.1484" xlink:type="simple">https://doi.org/10.1002/widm.1484</ext-link>.</mixed-citation>
</ref>
<ref id="j_infor601_ref_003">
<mixed-citation publication-type="journal"><string-name><surname>Chicco</surname>, <given-names>D.</given-names></string-name>, <string-name><surname>Warrens</surname>, <given-names>M.J.</given-names></string-name>, <string-name><surname>Jurman</surname>, <given-names>G.</given-names></string-name> (<year>2021</year>). <article-title>The Matthews Correlation Coefficient (MCC) is more informative than Cohen’s Kappa and brier score in binary classification assessment</article-title>. <source>IEEE Access</source>, <volume>9</volume>, <fpage>78368</fpage>–<lpage>78381</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1109/ACCESS.2021.3084050" xlink:type="simple">https://doi.org/10.1109/ACCESS.2021.3084050</ext-link>.</mixed-citation>
</ref>
<ref id="j_infor601_ref_004">
<mixed-citation publication-type="journal"><string-name><surname>Choi</surname>, <given-names>S.</given-names></string-name>, <string-name><surname>Cha</surname>, <given-names>S.</given-names></string-name>, <string-name><surname>Tappert</surname>, <given-names>C.C.</given-names></string-name> (<year>2010</year>). <article-title>A survey of binary similarity and distance measures</article-title>. <source>Journal of Systemics, Cybernetics and Informatics</source>, <volume>8</volume>(<issue>1</issue>), <fpage>43</fpage>–<lpage>48</lpage>.</mixed-citation>
</ref>
<ref id="j_infor601_ref_005">
<mixed-citation publication-type="journal"><string-name><surname>Demšar</surname>, <given-names>J.</given-names></string-name> (<year>2006</year>). <article-title>Statistical comparisons of classifiers over multiple data sets</article-title>. <source>Journal of Machine Learning Research</source>, <volume>7</volume>, <fpage>1</fpage>–<lpage>30</lpage>.</mixed-citation>
</ref>
<ref id="j_infor601_ref_006">
<mixed-citation publication-type="chapter"><string-name><surname>Feurer</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Hutter</surname>, <given-names>F.</given-names></string-name> (<year>2019</year>). <chapter-title>Hyperparameter optimization</chapter-title>. In: <string-name><surname>Hutter</surname>, <given-names>F.</given-names></string-name>, <string-name><surname>Kotthoff</surname>, <given-names>L.</given-names></string-name>, <string-name><surname>Vanschoren</surname>, <given-names>J.</given-names></string-name> (Eds.), <source>Automated Machine Learning: Methods, Systems, Challenges</source>. <publisher-name>Springer International Publishing</publisher-name>, pp. <fpage>3</fpage>–<lpage>33</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1007/978-3-030-05318-5_1" xlink:type="simple">https://doi.org/10.1007/978-3-030-05318-5_1</ext-link>.</mixed-citation>
</ref>
<ref id="j_infor601_ref_007">
<mixed-citation publication-type="chapter"><string-name><surname>Han</surname>, <given-names>J.</given-names></string-name>, <string-name><surname>Kamber</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Pei</surname>, <given-names>J.</given-names></string-name> (<year>2012</year>). <chapter-title>8 – Classification: Basic concepts</chapter-title>. In: <string-name><surname>Han</surname>, <given-names>J.</given-names></string-name>, <string-name><surname>Kamber</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Pei</surname>, <given-names>J.</given-names></string-name> (Eds.), <source>Data Mining <italic>(third edition)</italic></source>. <publisher-name>Morgan Kaufmann</publisher-name>, pp. <fpage>327</fpage>–<lpage>391</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1016/B978-0-12-381479-1.00008-3" xlink:type="simple">https://doi.org/10.1016/B978-0-12-381479-1.00008-3</ext-link>.</mixed-citation>
</ref>
<ref id="j_infor601_ref_008">
<mixed-citation publication-type="journal"><string-name><surname>Hubert</surname>, <given-names>L.</given-names></string-name> (<year>1977</year>). <article-title>Kappa revisited</article-title>. <source>Psychological Bulletin</source>, <volume>84</volume>(<issue>2</issue>), <fpage>289</fpage>–<lpage>297</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1037/0033-2909.84.2.289" xlink:type="simple">https://doi.org/10.1037/0033-2909.84.2.289</ext-link>.</mixed-citation>
</ref>
<ref id="j_infor601_ref_009">
<mixed-citation publication-type="journal"><string-name><surname>Janosi</surname>, <given-names>A.</given-names></string-name>, <string-name><surname>Steinbrunn</surname>, <given-names>W.</given-names></string-name>, <string-name><surname>Pfisterer</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Detrano</surname>, <given-names>R.</given-names></string-name> (<year>1988</year>). <article-title>Heart Disease [Dataset]</article-title>. <source>UCI Machine Learning Repository.</source> <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.24432/C52P4X" xlink:type="simple">https://doi.org/10.24432/C52P4X</ext-link>.</mixed-citation>
</ref>
<ref id="j_infor601_ref_010">
<mixed-citation publication-type="other"><string-name><surname>Kaggle</surname></string-name> (<year>2023</year>). Stroke prediction dataset [data retrieved from Kaggle]. <uri>https://www.kaggle.com/datasets/fedesoriano/stroke-prediction-dataset</uri>.</mixed-citation>
</ref>
<ref id="j_infor601_ref_011">
<mixed-citation publication-type="journal"><string-name><surname>Kuncheva</surname>, <given-names>L.I.</given-names></string-name>, <string-name><surname>Whitaker</surname>, <given-names>C.J.</given-names></string-name> (<year>2003</year>). <article-title>Measures of diversity in classifier ensembles and their relationship with the ensemble accuracy</article-title>. <source>Machine Learning</source>, <volume>51</volume>(<issue>2</issue>), <fpage>181</fpage>–<lpage>207</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1023/A:1022859003006" xlink:type="simple">https://doi.org/10.1023/A:1022859003006</ext-link>.</mixed-citation>
</ref>
<ref id="j_infor601_ref_012">
<mixed-citation publication-type="other"><string-name><surname>Labatut</surname>, <given-names>V.</given-names></string-name>, <string-name><surname>Cherifi</surname>, <given-names>H.</given-names></string-name> (<year>2012</year>). Accuracy measures for the comparison of classifiers. arXiv preprint. arXiv:<ext-link ext-link-type="uri" xlink:href="https://arxiv.org/abs/1207.3790">1207.3790</ext-link>.</mixed-citation>
</ref>
<ref id="j_infor601_ref_013">
<mixed-citation publication-type="journal"><string-name><surname>Liaw</surname>, <given-names>A.</given-names></string-name>, <string-name><surname>Wiener</surname>, <given-names>M.</given-names></string-name> (<year>2002</year>). <article-title>Classification and regression by randomforest</article-title>. <source>R News</source>, <volume>2</volume>(<issue>3</issue>), <fpage>18</fpage>–<lpage>22</lpage>. <comment><uri>https://CRAN.R-project.org/doc/Rnews/</uri></comment>.</mixed-citation>
</ref>
<ref id="j_infor601_ref_014">
<mixed-citation publication-type="chapter"><string-name><surname>Makhtar</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Neagu</surname>, <given-names>D.C.</given-names></string-name>, <string-name><surname>Ridley</surname>, <given-names>M.J.</given-names></string-name> (<year>2011</year>). <chapter-title>Binary classification models comparison: on the similarity of datasets and confusion matrix for predictive toxicology applications</chapter-title>. In: <string-name><surname>Böhm</surname>, <given-names>C.</given-names></string-name>, <string-name><surname>Khuri</surname>, <given-names>S.</given-names></string-name>, <string-name><surname>Lhotská</surname>, <given-names>L.</given-names></string-name>, <string-name><surname>Pisanti</surname>, <given-names>N.</given-names></string-name> (Eds.), <source>Information Technology in Bio- and Medical Informatics</source>. <publisher-name>Springer</publisher-name>, <publisher-loc>Berlin Heidelberg</publisher-loc>, pp. <fpage>108</fpage>–<lpage>122</lpage>.</mixed-citation>
</ref>
<ref id="j_infor601_ref_015">
<mixed-citation publication-type="journal"><string-name><surname>Margineantu</surname>, <given-names>D.D.</given-names></string-name>, <string-name><surname>Dietterich</surname>, <given-names>T.G.</given-names></string-name> (<year>1997</year>). <article-title>Pruning adaptive boosting</article-title>. <source>ICML</source>, <volume>97</volume>, <fpage>211</fpage>–<lpage>218</lpage>.</mixed-citation>
</ref>
<ref id="j_infor601_ref_016">
<mixed-citation publication-type="chapter"><string-name><surname>Narasimhamurthy</surname>, <given-names>A.</given-names></string-name> (<year>2005</year>). <chapter-title>Evaluation of diversity measures for binary classifier ensembles</chapter-title>. In: <string-name><surname>Oza</surname>, <given-names>N.C.</given-names></string-name>, <string-name><surname>Polikar</surname>, <given-names>R.</given-names></string-name>, <string-name><surname>Kittler</surname>, <given-names>J.</given-names></string-name>, <string-name><surname>Roli</surname>, <given-names>F.</given-names></string-name> (Eds.), <source>Multiple Classifier Systems</source>. <publisher-name>Springer</publisher-name>, <publisher-loc>Berlin Heidelberg</publisher-loc>, pp. <fpage>267</fpage>–<lpage>277</lpage>.</mixed-citation>
</ref>
<ref id="j_infor601_ref_017">
<mixed-citation publication-type="journal"><string-name><surname>Perišić</surname>, <given-names>A.</given-names></string-name>, <string-name><surname>Vanbelle</surname>, <given-names>S.</given-names></string-name> (<year>2024</year>). <article-title>Two-group k-adic similarity coefficients for binary classifiers</article-title>. <source>Journal of Classification</source>, <volume>41</volume>(<issue>2</issue>), <fpage>325</fpage>–<lpage>345</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1007/s00357-024-09498-8" xlink:type="simple">https://doi.org/10.1007/s00357-024-09498-8</ext-link>.</mixed-citation>
</ref>
<ref id="j_infor601_ref_018">
<mixed-citation publication-type="journal"><string-name><surname>Petrakos</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Atli Benediktsson</surname>, <given-names>J.</given-names></string-name>, <string-name><surname>Kanellopoulos</surname>, <given-names>I.</given-names></string-name> (<year>2001</year>). <article-title>The effect of classifier agreement on the accuracy of the combined classifier in decision level fusion</article-title>. <source>IEEE Transactions on Geoscience and Remote Sensing</source>, <volume>39</volume>(<issue>11</issue>), <fpage>2539</fpage>–<lpage>2546</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1109/36.964992" xlink:type="simple">https://doi.org/10.1109/36.964992</ext-link>.</mixed-citation>
</ref>
<ref id="j_infor601_ref_019">
<mixed-citation publication-type="other"><string-name><surname>Petričević</surname>, <given-names>R.B.</given-names></string-name> (<year>2023</year>). <italic>Primjena modificiranog k-adskog jaccardovog koeficijenta sličnosti za usporedbu dvaju skupova binarnih klasifikatora</italic>. Master’s thesis, University of Split, Faculty of Science in Split.</mixed-citation>
</ref>
<ref id="j_infor601_ref_020">
<mixed-citation publication-type="journal"><string-name><surname>Shirdel</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Di Mauro</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Liotta</surname>, <given-names>A.</given-names></string-name> (<year>2024</year>). <article-title>Worthiness benchmark: a novel concept for analyzing binary classification evaluation metrics</article-title>. <source>Information Sciences</source>, <volume>678</volume>, <elocation-id>120882</elocation-id>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1016/j.ins.2024.120882" xlink:type="simple">https://doi.org/10.1016/j.ins.2024.120882</ext-link>.</mixed-citation>
</ref>
<ref id="j_infor601_ref_021">
<mixed-citation publication-type="book"><string-name><surname>Sokal</surname>, <given-names>R.R.</given-names></string-name>, <string-name><surname>Sneath</surname>, <given-names>P.H.A.</given-names></string-name> (<year>1963</year>). <source>Principles of Numerical Taxonomy</source>. <publisher-name>W. H. Freeman; Company</publisher-name>.</mixed-citation>
</ref>
<ref id="j_infor601_ref_022">
<mixed-citation publication-type="journal"><string-name><surname>Strobl</surname>, <given-names>C.</given-names></string-name>, <string-name><surname>Boulesteix</surname>, <given-names>A.-L.</given-names></string-name>, <string-name><surname>Kneib</surname>, <given-names>T.</given-names></string-name>, <string-name><surname>Augustin</surname>, <given-names>T.</given-names></string-name>, <string-name><surname>Zeileis</surname>, <given-names>A.</given-names></string-name> (<year>2008</year>). <article-title>Conditional variable importance for random forests</article-title>. <source>BMC Bioinformatics</source>, <volume>9</volume> <elocation-id>307</elocation-id>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1186/1471-2105-9-307" xlink:type="simple">https://doi.org/10.1186/1471-2105-9-307</ext-link>.</mixed-citation>
</ref>
<ref id="j_infor601_ref_023">
<mixed-citation publication-type="journal"><string-name><surname>Strobl</surname>, <given-names>C.</given-names></string-name>, <string-name><surname>Boulesteix</surname>, <given-names>A.-L.</given-names></string-name>, <string-name><surname>Zeileis</surname>, <given-names>A.</given-names></string-name>, <string-name><surname>Hothorn</surname>, <given-names>T.</given-names></string-name> (<year>2007</year>). <article-title>Bias in random forest variable importance measures: illustrations, sources and a solution</article-title>. <source>BMC Bioinformatics</source>, <volume>8</volume>, <elocation-id>25</elocation-id>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1186/1471-2105-8-25" xlink:type="simple">https://doi.org/10.1186/1471-2105-8-25</ext-link>.</mixed-citation>
</ref>
<ref id="j_infor601_ref_024">
<mixed-citation publication-type="journal"><string-name><surname>Strobl</surname>, <given-names>C.</given-names></string-name>, <string-name><surname>Hothorn</surname>, <given-names>T.</given-names></string-name>, <string-name><surname>Zeileis</surname>, <given-names>A.</given-names></string-name> (<year>2009</year>). <article-title>Party on!</article-title> <source>The R Journal</source>, <volume>1</volume>(<issue>2</issue>), <fpage>14</fpage>–<lpage>17</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.32614/RJ-2009-013" xlink:type="simple">https://doi.org/10.32614/RJ-2009-013</ext-link>.</mixed-citation>
</ref>
<ref id="j_infor601_ref_025">
<mixed-citation publication-type="journal"><string-name><surname>Tang</surname>, <given-names>E.K.</given-names></string-name>, <string-name><surname>Suganthan</surname>, <given-names>P.N.</given-names></string-name>, <string-name><surname>Yao</surname>, <given-names>X.</given-names></string-name> (<year>2006</year>). <article-title>An analysis of diversity measures</article-title>. <source>Machine Learning</source>, <volume>65</volume>(<issue>1</issue>), <fpage>247</fpage>–<lpage>271</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1007/s10994-006-9449-2" xlink:type="simple">https://doi.org/10.1007/s10994-006-9449-2</ext-link>.</mixed-citation>
</ref>
<ref id="j_infor601_ref_026">
<mixed-citation publication-type="journal"><string-name><surname>Tsymbal</surname>, <given-names>A.</given-names></string-name>, <string-name><surname>Pechenizkiy</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Cunningham</surname>, <given-names>P.</given-names></string-name> (<year>2005</year>). <article-title>Diversity in search strategies for ensemble feature selection</article-title>. <source>Information Fusion</source>, <volume>6</volume>(<issue>1</issue>), <fpage>83</fpage>–<lpage>98</lpage>.</mixed-citation>
</ref>
<ref id="j_infor601_ref_027">
<mixed-citation publication-type="journal"><string-name><surname>Vanbelle</surname>, <given-names>S.</given-names></string-name>, <string-name><surname>Albert</surname>, <given-names>A.</given-names></string-name> (<year>2009</year>). <article-title>Agreement between two independent groups of raters</article-title>. <source>Psychometrika</source>, <volume>74</volume>, <fpage>477</fpage>–<lpage>491</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1007/S11336-009-9116-1" xlink:type="simple">https://doi.org/10.1007/S11336-009-9116-1</ext-link>.</mixed-citation>
</ref>
<ref id="j_infor601_ref_028">
<mixed-citation publication-type="journal"><string-name><surname>Wang</surname>, <given-names>J.</given-names></string-name>, <string-name><surname>Wang</surname>, <given-names>L.</given-names></string-name>, <string-name><surname>Zheng</surname>, <given-names>Y.</given-names></string-name>, <string-name><surname>Yeh</surname>, <given-names>C.-C.M.</given-names></string-name>, <string-name><surname>Jain</surname>, <given-names>S.</given-names></string-name>, <string-name><surname>Zhang</surname>, <given-names>W.</given-names></string-name> (<year>2023</year>). <article-title>Learning-from-disagreement: a model comparison and visual analytics framework</article-title>. <source>IEEE Transactions on Visualization &amp; Computer Graphics</source>, <volume>29</volume>(<issue>09</issue>), <fpage>3809</fpage>–<lpage>3825</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1109/TVCG.2022.3172107" xlink:type="simple">https://doi.org/10.1109/TVCG.2022.3172107</ext-link>.</mixed-citation>
</ref>
<ref id="j_infor601_ref_029">
<mixed-citation publication-type="journal"><string-name><surname>Warrens</surname>, <given-names>M.J.</given-names></string-name> (<year>2009</year>). <article-title>k-adic similarity coefficients for binary (presence/absence) data</article-title>. <source>Journal of Classification</source>, <volume>26</volume>, <fpage>227</fpage>–<lpage>245</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1007/s00357-009-9032-1" xlink:type="simple">https://doi.org/10.1007/s00357-009-9032-1</ext-link>.</mixed-citation>
</ref>
<ref id="j_infor601_ref_030">
<mixed-citation publication-type="journal"><string-name><surname>Wood</surname>, <given-names>D.</given-names></string-name>, <string-name><surname>Mu</surname>, <given-names>T.</given-names></string-name>, <string-name><surname>Webb</surname>, <given-names>A.M.</given-names></string-name>, <string-name><surname>Reeve</surname>, <given-names>H.W.</given-names></string-name>, <string-name><surname>Lujan</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Brown</surname>, <given-names>G.</given-names></string-name> (<year>2023</year>). <article-title>A unified theory of diversity in ensemble learning</article-title>. <source>Journal of Machine Learning Research</source>, <volume>24</volume>(<issue>359</issue>), <fpage>1</fpage>–<lpage>49</lpage>.</mixed-citation>
</ref>
<ref id="j_infor601_ref_031">
<mixed-citation publication-type="chapter"><string-name><surname>Zouari</surname>, <given-names>H.</given-names></string-name>, <string-name><surname>Heutte</surname>, <given-names>L.</given-names></string-name>, <string-name><surname>Lecourtier</surname>, <given-names>Y.</given-names></string-name> (<year>2005</year>). <chapter-title>Using diversity measure in building classifier ensembles for combination method analysis</chapter-title>. In: <string-name><surname>Kurzyński</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Puchała</surname>, <given-names>E.</given-names></string-name>, <string-name><surname>Woźniak</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>żołnierek</surname>, <given-names>A.</given-names></string-name> (Eds.), <source>Computer Recognition Systems</source>. <publisher-name>Springer</publisher-name>, <publisher-loc>Berlin Heidelberg</publisher-loc>, pp. <fpage>337</fpage>–<lpage>344</lpage>.</mixed-citation>
</ref>
</ref-list>
</back>
</article>
