<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article
  PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD with MathML3 v1.2 20190208//EN" "JATS-journalpublishing1-mathml3.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xmlns:ali="http://www.niso.org/schemas/ali/1.0/" article-type="research-article" dtd-version="1.2" xml:lang="en">
<front>
<journal-meta><journal-id journal-id-type="publisher-id">METH</journal-id><journal-id journal-id-type="nlm-ta">Methodology</journal-id>
<journal-title-group>
<journal-title>Methodology</journal-title><abbrev-journal-title abbrev-type="pubmed">Methodology</abbrev-journal-title>
</journal-title-group>
<issn pub-type="ppub">1614-1881</issn>
<issn pub-type="epub">1614-2241</issn>
<publisher><publisher-name>PsychOpen</publisher-name></publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="publisher-id">meth.14931</article-id>
<article-id pub-id-type="doi">10.5964/meth.14931</article-id>
<article-categories>
<subj-group subj-group-type="heading"><subject>Original Article</subject></subj-group>

<subj-group subj-group-type="badge">
<subject>Code</subject>
<subject>Materials</subject>
</subj-group>

</article-categories>
<title-group>
<article-title>Selecting the Number of Clusters in Mixture Multigroup Structural Equation Modeling</article-title>
	<alt-title alt-title-type="right-running">Model Selection in MMG-SEM</alt-title>
	<alt-title specific-use="APA-reference-style" xml:lang="en">Selecting the number of clusters in Mixture Multigroup Structural Equation Modeling</alt-title>
</title-group>
<contrib-group content-type="authors">
<contrib id="author-1" contrib-type="author" corresp="yes"><contrib-id contrib-id-type="orcid" authenticated="false">https://orcid.org/0000-0002-2480-8771</contrib-id><name name-style="western"><surname>Perez Alonso</surname><given-names>Andres F.</given-names></name><xref ref-type="corresp" rid="cor1">*</xref><xref ref-type="aff" rid="aff1">1</xref><xref ref-type="aff" rid="aff2">2</xref></contrib>
<contrib id="author-2" contrib-type="author"><contrib-id contrib-id-type="orcid" authenticated="false">https://orcid.org/0000-0001-9053-9330</contrib-id><name name-style="western"><surname>Vermunt</surname><given-names>Jeroen K.</given-names></name><xref ref-type="aff" rid="aff1">1</xref></contrib>
<contrib id="author-3" contrib-type="author"><contrib-id contrib-id-type="orcid" authenticated="false">https://orcid.org/0000-0002-4129-4477</contrib-id><name name-style="western"><surname>Rosseel</surname><given-names>Yves</given-names></name><xref ref-type="aff" rid="aff3">3</xref></contrib>
<contrib id="author-4" contrib-type="author"><contrib-id contrib-id-type="orcid" authenticated="false">https://orcid.org/0000-0002-0299-0648</contrib-id><name name-style="western"><surname>Roover</surname><given-names>Kim De</given-names></name><xref ref-type="aff" rid="aff1">1</xref><xref ref-type="aff" rid="aff2">2</xref></contrib>
<contrib contrib-type="editor">
<name>
<surname>Estrada</surname>
<given-names>Eduardo</given-names>
</name>
<xref ref-type="aff" rid="aff4"/>
</contrib>
	<aff id="aff1"><label>1</label><institution>Department of Methodology and Statistics, Tilburg University, Tilburg</institution>, <country country="NL">the Netherlands</country></aff>
	<aff id="aff2"><label>2</label><institution>Research Group of Quantitative Psychology and Individual Differences, KU Leuven, Leuven</institution>, <country country="BE">Belgium</country></aff>
	<aff id="aff3"><label>3</label><institution>Department of Data Analysis, Ghent University, Ghent</institution>, <country country="BE">Belgium</country></aff>
	<aff id="aff4">Autonomous University of Madrid, Madrid, <country>Spain</country></aff>
</contrib-group>
<author-notes>
	<corresp id="cor1"><label>*</label>Department of Methodology and Statistics, Tilburg University, PO Box 90153 5000 LE, Tilburg, the Netherlands. <email xlink:href="A.F.PerezAlonso@tilburguniversity.edu">A.F.PerezAlonso@tilburguniversity.edu</email></corresp>
</author-notes>
<pub-date pub-type="epub"><day>31</day><month>03</month><year>2025</year></pub-date>
<pub-date pub-type="collection" publication-format="electronic"><year>2025</year></pub-date>
<volume>21</volume>
<issue>1</issue>
<fpage>1</fpage>
<lpage>26</lpage>
<history>
<date date-type="received">
<day>25</day>
<month>06</month>
<year>2024</year>
</date>
<date date-type="accepted">
<day>04</day>
<month>02</month>
<year>2025</year>
</date>
</history>
	
<permissions><copyright-year>2025</copyright-year><copyright-holder>Perez Alonso, Vermunt, Rosseel, &amp; Roover</copyright-holder><license license-type="open-access" specific-use="CC BY 4.0" xlink:href="https://creativecommons.org/licenses/by/4.0/"><ali:license_ref>https://creativecommons.org/licenses/by/4.0/</ali:license_ref><license-p>This is an open-access article distributed under the terms of the Creative Commons Attribution (CC BY) 4.0 License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.</license-p></license></permissions>	
	
<abstract>
	<p>Behavioral scientists often use Multigroup Structural Equation Modeling (MG-SEM) to compare groups in terms of their latent variables (LVs) relations — also called 'structural relations’. Since LVs are measured indirectly, measurement invariance must be evaluated before comparing structural relations. To efficiently compare many groups, the recently proposed Mixture MG-SEM (MMG-SEM) clusters groups based on their structural relations while accounting for measurement (non-)invariance. MMG-SEM requires the user to select the optimal number of clusters for the data at hand. Various approaches address this problem, but no definitive answer exists on which is best. This paper aims to find the best-performing model selection approach for MMG-SEM through a simulation study by comparing five information criteria and the convex hull procedure and including empirically realistic conditions affecting the clusters’ separability. No universally best measure was found, but based on our results, we recommend using the convex hull combined with another measure (e.g., AIC) when selecting the number of clusters.</p>
</abstract>
<kwd-group kwd-group-type="author"><kwd>model selection</kwd><kwd>mixture modeling</kwd><kwd>structural relations</kwd><kwd>structural equation modeling</kwd></kwd-group>

</article-meta>
</front>
<body>
	<sec sec-type="intro" id="intro"><title/>
<p>Comparing relations between unobservable or ‘latent’ variables (e.g., attitudes, emotions) across many groups is common in behavioral sciences. For instance, <xref ref-type="bibr" rid="r26">Mayerl and Best (2019)</xref> studied how environmental attitudes related to environmental behavior in 30 countries. Structural Equation Modeling (SEM; <xref ref-type="bibr" rid="r6">Bollen, 1989</xref>) allows estimating regression coefficients for relations among latent variables (LV) based on the covariances of their observed indicators, such as questionnaire items. In SEM, the regression coefficients are also called ‘structural relations’ and LVs are called ‘factors’.</p>
<p>Multigroup SEM (MG-SEM) and Multilevel SEM (ML-SEM), are commonly used to compare structural relations across groups (e.g., <xref ref-type="bibr" rid="r26">Mayerl &amp; Best, 2019</xref>). To pinpoint differences and similarities, these approaches require pairwise comparisons of the group-specific values for the structural relations, which is a complex and daunting task when many groups are involved. For instance, for 30 groups, this would entail 435 pairwise comparisons per parameter. In ML-SEM, the group-specific parameter values are derived from random effects (<xref ref-type="bibr" rid="r20">Hox et al., 2017</xref>).</p>
<p>An intuitive solution is to find subsets of groups that share the same relations between factors using mixture modeling (<xref ref-type="bibr" rid="r28">McLachlan et al., 2019</xref>). However, before identifying such ‘latent classes’ or ‘clusters’, we must remember that the factors are indirectly measured via questionnaire items. Before comparing structural relations between groups, we must ensure that the measurement of the factors is the same across groups or, in other words, that ‘measurement invariance’ (MI; <xref ref-type="bibr" rid="r29">Meredith, 1993</xref>) holds. This implies evaluating whether the measurement model (MM) — indicating which items measure which factors and to what extent — is invariant across groups. The MI assumption can be evaluated at several levels, focusing on different MM parameters (<xref ref-type="bibr" rid="r38">Vandenberg &amp; Lance, 2000</xref>). In case of many groups, invariance often does not hold for all the MM parameters. To compare structural relations across groups, the equality of the so-called ‘factor loadings’ (i.e., the item-factor relations) must hold, which is called metric invariance<xref ref-type="fn" rid="fn1"><sup>1</sup></xref><fn id="fn1"><label>1</label>
<p>When full metric invariance does not hold (i.e., all loadings equal), partial metric invariance (<xref ref-type="bibr" rid="r9">Byrne et al., 1989</xref>) can be pursued, where some of the loadings are allowed to be different across groups.</p></fn>. Other higher-level MM differences are inconsequential for comparing structural relations if they are included in the model (<xref ref-type="bibr" rid="r13">Chen, 2008</xref>; <xref ref-type="bibr" rid="r19">Guenole &amp; Brown, 2014</xref>).</p>
<p>When looking for clusters of groups with equal structural relations, capturing the MM differences — or ‘measurement non-invariances’ — with group-specific parameters is important. Many existing mixture SEM methods (e.g., <xref ref-type="bibr" rid="r22">Kim et al., 2016</xref>; <xref ref-type="bibr" rid="r39">Vermunt &amp; Magidson, 2005</xref>) force all parameters to be equal within a cluster. This implies that MM parameters can either be specified as invariant across all groups (i.e., ignoring measurement non-invariances) or as cluster-specific (i.e., enforcing MI within each cluster but not across clusters). Such mixture SEM methods capture clusters of groups with the same structural relations as well as the same MM parameters, and fail to disentangle the differences of interest (i.e., in the structural relations) from those not of interest (i.e., in the MM).</p>
<p>To effectively capture clusters of groups with equivalent structural relations while simultaneously accounting for measurement non-invariances, <xref ref-type="bibr" rid="r31">Perez Alonso et al. (2024)</xref> proposed Mixture Multigroup SEM (MMG-SEM), which combines MG-SEM with mixture clustering. Specifically, it combines cluster-specific structural relations with measurement parameters that are partially group-specific, so that the clustering of the groups focuses only on the structural relations, which are of interest to the research question. Note that, essentially, MMG-SEM comprises two different types of LVs: (1) the continuous LVs measured by items at the individual-level, and (2) a categorical LV for the clusters at the group-level.</p>
<p>By gathering groups with equal structural relations in a cluster, MMG-SEM reduces the number of pairwise comparisons needed to pinpoint which relations differ among groups. However, it also introduces a problem inherent to mixture models; that is, for each data set, the appropriate number of clusters must be determined. In empirical research, the ‘true’ number of clusters is unknown, and the selection of the number of clusters is an important challenge. When too few clusters are selected, one fails to detect potentially interesting differences in the structural relations and, when too many clusters are retained, one ends up with an overly complex model. <xref ref-type="bibr" rid="r31">Perez Alonso et al. (2024)</xref> showed that MMG-SEM performs well when the correct number of clusters is specified, but did not address the model selection problem. In their empirical application, they applied a model selection approach recommended for related mixture methods (e.g., <xref ref-type="bibr" rid="r16">De Roover et al., 2022</xref>; <xref ref-type="bibr" rid="r24">Lukočiene et al., 2010</xref>; <xref ref-type="bibr" rid="r25">Lukočiene &amp; Vermunt, 2009</xref>), but they did not evaluate how commonly used model selection approaches perform for MMG-SEM in different conditions or which approach is best for MMG-SEM.</p>
<p>Several approaches to address the model selection problem are available (see <xref ref-type="bibr" rid="r2">Akogul &amp; Erisoglu, 2016</xref>). For instance, the Bayesian Information Criterion (BIC; <xref ref-type="bibr" rid="r35">Schwarz, 1978</xref>) and Akaike Information Criterion (AIC; <xref ref-type="bibr" rid="r1">Akaike, 1974</xref>) integrate model fit and a penalty based on model complexity. Hence, a model that minimizes the criteria is assumed to have a good balance between model fit and parsimony. Another way of finding this balance is using the Convex Hull method (<xref ref-type="bibr" rid="r11">Ceulemans &amp; Kiers, 2006</xref>), which is a generalized scree test. Alternatively, the Integrated Completed Likelihood (ICL; <xref ref-type="bibr" rid="r5">Biernacki et al., 2000</xref>) also considers the cluster separation; that is, it penalizes models that offer poorly-defined clusters (i.e., clusters that are too similar). This aligns with the fact that substantive researchers likely regard minor differences in structural relations to be trivial.</p>
<p>Numerous simulation studies have compared different model selection methods (e.g., <xref ref-type="bibr" rid="r2">Akogul &amp; Erisoglu, 2016</xref>; <xref ref-type="bibr" rid="r16">De Roover et al., 2022</xref>; <xref ref-type="bibr" rid="r24">Lukočiene et al., 2010</xref>; <xref ref-type="bibr" rid="r25">Lukočiene &amp; Vermunt, 2009</xref>; <xref ref-type="bibr" rid="r30">Nylund et al., 2007</xref>), showing different results depending on the conditions and mixture models evaluated. For instance, for mixture models combined with factor analysis, <xref ref-type="bibr" rid="r8">Bulteel et al. (2013)</xref> and <xref ref-type="bibr" rid="r15">De Roover (2021)</xref> found that BIC and Convex Hull outperformed AIC. In the context of latent class analysis, <xref ref-type="bibr" rid="r25">Lukočiene and Vermunt (2009)</xref> found that <inline-formula id="ieqn-1"><mml:math id="mml-ieqn-1"><mml:msub><mml:mrow><mml:mi mathvariant="normal">A</mml:mi><mml:mi mathvariant="normal">I</mml:mi><mml:mi mathvariant="normal">C</mml:mi></mml:mrow><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msub></mml:math></inline-formula> (i.e., a modified AIC with a larger penalty) performed better than other model selection methods.</p>
<p>These contradictory results emphasize the importance of evaluating and comparing model selection approaches for MMG-SEM specifically, which is the aim of this paper. By means of a simulation study, we will compare different approaches in conditions that mimic the ones found in social sciences. For instance, in empirical data, it is likely that certain groups have very similar — but not identical — regression parameters. Gathering these groups in the same cluster may still be desirable, for the sake of parsimony, and because researchers are often not interested in such trivial differences. Therefore, the simulated conditions will include different levels of small differences in structural relations <italic>within</italic> a cluster, to evaluate how this affects the model selection.</p>
<p>The remainder of this paper is organized as follows: MMG-SEM and relevant model selection methods are described in the Method section. A Simulation Study then evaluates the performance of the model selection methods in the context of MMG-SEM. The paper concludes with a Discussion section highlighting the most relevant results and limitations of the study.</p></sec>
<sec sec-type="methods"><title>Method</title>
<sec><title>Mixture Multigroup Structural Equation Modeling</title>
<p>Mixture Multigroup Structural Equation Modeling (MMG-SEM; <xref ref-type="bibr" rid="r31">Perez Alonso et al., 2024</xref>) combines mixture modeling with MG-SEM. In general, the mixture multigroup approach (<xref ref-type="bibr" rid="r15">De Roover, 2021</xref>; <xref ref-type="bibr" rid="r16">De Roover et al., 2022</xref>), aims to find a clustering that focuses on specific parameters of interest. In MMG-SEM, the clustering focuses on the structural relations, while MM differences are accounted for by group-specific parameters, so they do not affect the clustering.</p>
<p id="S2.SS1.Px1.p2">For its estimation, <xref ref-type="bibr" rid="r31">Perez Alonso et al. (2024)</xref> used the ‘Structural-After-Measurement’ (SAM; <xref ref-type="bibr" rid="r33">Rosseel &amp; Loh, 2022</xref>) approach, which estimates a SEM model in two steps. In the first step, the MM is estimated, whereas the structural model (SM; including the structural relations) is estimated in the second step. The estimation of MMG-SEM is briefly described below (for more details, see <xref ref-type="bibr" rid="r31">Perez Alonso et al., 2024</xref>).</p>
<sec><title>Step 1: Measurement Model</title>
<p>The MM defines how the LVs are measured; that is, which items measure which factor and to what extent. When studying multiple groups (e.g., countries), the MM is often estimated using Multigroup Confirmatory Factor Analysis (MG-CFA). If we consider individuals <inline-formula id="ieqn-2"><mml:math id="mml-ieqn-2"><mml:msub><mml:mi>n</mml:mi><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mo>…</mml:mo><mml:mo>,</mml:mo><mml:msub><mml:mi>N</mml:mi><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> within groups <inline-formula id="ieqn-3"><mml:math id="mml-ieqn-3"><mml:mi>g</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mo>…</mml:mo><mml:mo>,</mml:mo><mml:mi>G</mml:mi></mml:math></inline-formula>, items <inline-formula id="ieqn-4"><mml:math id="mml-ieqn-4"><mml:mi>j</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mo>…</mml:mo><mml:mo>,</mml:mo><mml:mi>J</mml:mi></mml:math></inline-formula>, and factors <inline-formula id="ieqn-5"><mml:math id="mml-ieqn-5"><mml:mi>q</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mo>…</mml:mo><mml:mo>,</mml:mo><mml:mi>Q</mml:mi></mml:math></inline-formula>, MG-CFA defines the vector of observed scores <inline-formula id="ieqn-6"><mml:math id="mml-ieqn-6"><mml:msub><mml:mrow><mml:mi mathvariant="bold">x</mml:mi></mml:mrow><mml:mrow><mml:msub><mml:mi>n</mml:mi><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:msub></mml:math></inline-formula> of individual <italic>n</italic><sub><italic>g</italic></sub> as follows 
<disp-formula id="eqn-1"><label>1</label><mml:math id="mml-eqn-1" display="block"><mml:msub><mml:mrow><mml:mtext>x</mml:mtext></mml:mrow><mml:mrow><mml:msub><mml:mi>n</mml:mi><mml:mi>g</mml:mi></mml:msub></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:msub><mml:mi mathvariant="bold-italic">τ</mml:mi><mml:mi>g</mml:mi></mml:msub><mml:mo>+</mml:mo><mml:msub><mml:mi mathvariant="bold">Λ</mml:mi><mml:mi>g</mml:mi></mml:msub><mml:msub><mml:mi mathvariant="bold-italic">η</mml:mi><mml:mrow><mml:msub><mml:mi>n</mml:mi><mml:mi>g</mml:mi></mml:msub></mml:mrow></mml:msub><mml:mo>+</mml:mo><mml:msub><mml:mi mathvariant="bold-italic">ϵ</mml:mi><mml:mrow><mml:msub><mml:mi>n</mml:mi><mml:mi>g</mml:mi></mml:msub></mml:mrow></mml:msub><mml:mo>,</mml:mo></mml:math></disp-formula>
</p>
<p>where <inline-formula id="ieqn-7"><mml:math id="mml-ieqn-7"><mml:msub><mml:mi mathvariant="bold-italic">τ</mml:mi><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> is a <italic>J</italic>-dimensional vector of group-specific intercepts, <inline-formula id="ieqn-8"><mml:math id="mml-ieqn-8"><mml:msub><mml:mi mathvariant="bold">Λ</mml:mi><mml:mrow><mml:mi mathvariant="bold-italic">g</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> denotes a <italic>J </italic>× <italic>Q</italic> matrix of group-specific factor loadings, <inline-formula id="ieqn-9"><mml:math id="mml-ieqn-9"><mml:msub><mml:mi mathvariant="bold-italic">η</mml:mi><mml:mrow><mml:msub><mml:mi>n</mml:mi><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:msub></mml:math></inline-formula> is a <italic>Q</italic>-dimensional random vector of factor scores, and <inline-formula id="ieqn-10"><mml:math id="mml-ieqn-10"><mml:msub><mml:mi mathvariant="bold-italic">ϵ</mml:mi><mml:mrow><mml:msub><mml:mi>n</mml:mi><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:msub></mml:math></inline-formula> is a <italic>J</italic>-dimensional random vector of residuals. Note that MG-CFA imposes a pattern of zero and non-zero loadings on <inline-formula id="ieqn-11"><mml:math id="mml-ieqn-11"><mml:msub><mml:mi mathvariant="bold">Λ</mml:mi><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, and that, in this paper, we center the observed variables per group to remove the mean structure, which is equivalent to estimating the group-specific <inline-formula id="ieqn-12"><mml:math id="mml-ieqn-12"><mml:msub><mml:mi mathvariant="bold-italic">τ</mml:mi><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, but computationally more efficient. We assume that: (1) <inline-formula id="ieqn-13"><mml:math id="mml-ieqn-13"><mml:msub><mml:mi mathvariant="bold-italic">η</mml:mi><mml:mrow><mml:msub><mml:mi>n</mml:mi><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:msub></mml:math></inline-formula> is distributed according to a multivariate normal distribution <inline-formula id="ieqn-14"><mml:math id="mml-ieqn-14"><mml:mi>M</mml:mi><mml:mi>V</mml:mi><mml:mi>N</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi mathvariant="bold-italic">α</mml:mi><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mi mathvariant="bold">Φ</mml:mi><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula>, where <inline-formula id="ieqn-15"><mml:math id="mml-ieqn-15"><mml:msub><mml:mi mathvariant="bold-italic">α</mml:mi><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> and <inline-formula id="ieqn-16"><mml:math id="mml-ieqn-16"><mml:msub><mml:mi mathvariant="bold">Φ</mml:mi><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> are the mean vector and covariance matrix of the factors, respectively, and (2) <inline-formula id="ieqn-17"><mml:math id="mml-ieqn-17"><mml:msub><mml:mi mathvariant="bold-italic">ϵ</mml:mi><mml:mrow><mml:msub><mml:mi>n</mml:mi><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:msub></mml:math></inline-formula> is distributed according to <inline-formula id="ieqn-18"><mml:math id="mml-ieqn-18"><mml:mi>M</mml:mi><mml:mi>V</mml:mi><mml:mi>N</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mn>0</mml:mn><mml:mo>,</mml:mo><mml:msub><mml:mi mathvariant="bold">Θ</mml:mi><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula>, where <inline-formula id="ieqn-19"><mml:math id="mml-ieqn-19"><mml:msub><mml:mi mathvariant="bold">Θ</mml:mi><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> is the covariance matrix of the residuals in Group <italic>g</italic>, which is usually assumed to be diagonal. If we assume that <inline-formula id="ieqn-20"><mml:math id="mml-ieqn-20"><mml:mrow><mml:mi mathvariant="normal">C</mml:mi><mml:mi mathvariant="normal">o</mml:mi><mml:mi mathvariant="normal">v</mml:mi></mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi mathvariant="bold-italic">η</mml:mi><mml:mrow><mml:msub><mml:mi>n</mml:mi><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mi mathvariant="bold-italic">ϵ</mml:mi><mml:mrow><mml:msub><mml:mi>n</mml:mi><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:msub><mml:mo stretchy="false">)</mml:mo><mml:mo>=</mml:mo><mml:mn mathvariant="bold">0</mml:mn></mml:math></inline-formula>, the model-implied covariance matrix of Group <italic>g</italic> is given by 
<disp-formula id="eqn-2"><label>2</label><mml:math id="mml-eqn-2" display="block"><mml:msub><mml:mi mathvariant="bold">Σ</mml:mi><mml:mi>g</mml:mi></mml:msub><mml:mo>=</mml:mo><mml:msub><mml:mi mathvariant="bold">Λ</mml:mi><mml:mi>g</mml:mi></mml:msub><mml:msub><mml:mi mathvariant="bold">Φ</mml:mi><mml:mi>g</mml:mi></mml:msub><mml:msubsup><mml:mi mathvariant="bold">Λ</mml:mi><mml:mi>g</mml:mi><mml:mo>′</mml:mo></mml:msubsup><mml:mo>+</mml:mo><mml:msub><mml:mi mathvariant="bold">Θ</mml:mi><mml:mi>g</mml:mi></mml:msub><mml:mo>.</mml:mo></mml:math></disp-formula>
</p>
<p>In the context of MMG-SEM, one must evaluate measurement invariance (MI) before clustering groups on structural relations. As mentioned in the <xref ref-type="sec" rid="intro	">Introduction</xref>, MI can be evaluated at different levels by focusing on different parameters. First, one must evaluate configural invariance by testing if the model in <xref ref-type="disp-formula" rid="eqn-1">Equation 1</xref> holds across groups. If the model’s fit is satisfactory (<xref ref-type="bibr" rid="r12">Chen, 2007</xref>), one can assume that the same items measure the same factors across groups. Second, one must test for metric invariance by constraining the non-zero loadings in <inline-formula id="ieqn-21"><mml:math id="mml-ieqn-21"><mml:msub><mml:mi mathvariant="bold">Λ</mml:mi><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> to be the same across groups. If imposing <inline-formula id="ieqn-22"><mml:math id="mml-ieqn-22"><mml:msub><mml:mi mathvariant="bold">Λ</mml:mi><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mi mathvariant="bold">Λ</mml:mi></mml:math></inline-formula> does not significantly worsen the model fit (<xref ref-type="bibr" rid="r34">Rutkowski &amp; Svetina, 2014</xref>), full metric invariance holds. Metric invariance must hold, at least partially, for valid comparisons of the structural relations<xref ref-type="fn" rid="fn2"><sup>2</sup></xref><fn id="fn2"><label>2</label>
<p>For more information about the remaining MI levels, please see <xref ref-type="bibr" rid="r38">Vandenberg and Lance (2000)</xref>.</p></fn>. Therefore, the remaining MM parameters (i.e., <inline-formula id="ieqn-23"><mml:math id="mml-ieqn-23"><mml:msub><mml:mi mathvariant="bold-italic">τ</mml:mi><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, <inline-formula id="ieqn-24"><mml:math id="mml-ieqn-24"><mml:msub><mml:mi mathvariant="bold">Θ</mml:mi><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, and potentially some loadings) are allowed to be group-specific in MMG-SEM. If full metric invariance holds, the model-implied covariance matrix of Group <italic>g</italic> in MMG-SEM’s first step is 
<disp-formula id="eqn-3"><label>3</label><mml:math id="mml-eqn-3" display="block"><mml:msub><mml:mi mathvariant="bold">Σ</mml:mi><mml:mi>g</mml:mi></mml:msub><mml:mo>=</mml:mo><mml:mi mathvariant="bold">Λ</mml:mi><mml:msub><mml:mi mathvariant="bold">Φ</mml:mi><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub><mml:msup><mml:mi mathvariant="bold">Λ</mml:mi><mml:mo>′</mml:mo></mml:msup><mml:mo>+</mml:mo><mml:msub><mml:mi mathvariant="bold">Θ</mml:mi><mml:mi>g</mml:mi></mml:msub><mml:mo>.</mml:mo></mml:math></disp-formula>
</p>
<p>The MM is fitted by minimizing the difference between the model-implied covariance matrix <inline-formula id="ieqn-25"><mml:math id="mml-ieqn-25"><mml:msub><mml:mi mathvariant="bold">Σ</mml:mi><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> and the observed covariance matrix <inline-formula id="ieqn-26"><mml:math id="mml-ieqn-26"><mml:mrow><mml:mi mathvariant="bold">S</mml:mi></mml:mrow></mml:math></inline-formula>. The group-specific factor covariance matrices <inline-formula id="ieqn-27"><mml:math id="mml-ieqn-27"><mml:msub><mml:mi mathvariant="bold">Φ</mml:mi><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> from <xref ref-type="disp-formula" rid="eqn-3">Equation 3</xref> are the input for MMG-SEM’s second step. To avoid confusion in the notation of the remaining text, the covariance matrices <inline-formula id="ieqn-28"><mml:math id="mml-ieqn-28"><mml:msub><mml:mi mathvariant="bold">Φ</mml:mi><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> from Step 1 will have a superscript <italic>s</italic>1 (i.e., <inline-formula id="ieqn-29"><mml:math id="mml-ieqn-29"><mml:msubsup><mml:mi mathvariant="bold">Φ</mml:mi><mml:mrow><mml:mi>g</mml:mi></mml:mrow><mml:mrow><mml:mi>s</mml:mi><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:math></inline-formula>).</p></sec>
<sec><title>Step 2: Structural Model</title>
<p>In the second step, MMG-SEM estimates the SM and performs the mixture clustering based on the structural relations. Note that MMG-SEM operates at the group-level; that is, it finds clusters of groups instead of clusters of observations. The SM, which defines how the LVs are related, is conditional on the membership of Group <italic>g</italic> to Cluster <italic>k</italic>, denoted as <italic>z</italic><sub><italic>gk</italic></sub>, which takes on a value of 1 or 0. Note that the true cluster memberships <italic>z</italic><sub><italic>gk</italic></sub> are unknown and that the estimated <inline-formula id="ieqn-30"><mml:math id="mml-ieqn-30"><mml:msub><mml:mrow><mml:mover><mml:mi>z</mml:mi><mml:mo stretchy="false">^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>g</mml:mi><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> is a probability ranging from 0 to 1. Formally, the model-implied factor covariance matrix <inline-formula id="ieqn-31"><mml:math id="mml-ieqn-31"><mml:msub><mml:mi mathvariant="bold">Φ</mml:mi><mml:mrow><mml:mi>g</mml:mi><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> is defined as: 
<disp-formula id="eqn-4"><label>4</label><mml:math id="mml-eqn-4" display="block"><mml:mo stretchy="false">[</mml:mo><mml:msub><mml:mi mathvariant="bold">Φ</mml:mi><mml:mrow><mml:mi>g</mml:mi><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">|</mml:mo></mml:mrow><mml:msub><mml:mi>z</mml:mi><mml:mrow><mml:mi>g</mml:mi><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mn>1</mml:mn><mml:mo stretchy="false">]</mml:mo><mml:mo>=</mml:mo><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mtext>I</mml:mtext></mml:mrow><mml:mo>−</mml:mo><mml:msub><mml:mrow><mml:mtext>B</mml:mtext></mml:mrow><mml:mi>k</mml:mi></mml:msub><mml:msup><mml:mo stretchy="false">)</mml:mo><mml:mrow><mml:mo>−</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msup><mml:msub><mml:mi mathvariant="bold">Ψ</mml:mi><mml:mrow><mml:mi>g</mml:mi><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mtext>I</mml:mtext></mml:mrow><mml:mo>−</mml:mo><mml:msub><mml:mrow><mml:mtext>B</mml:mtext></mml:mrow><mml:mi>k</mml:mi></mml:msub><mml:msup><mml:mo stretchy="false">)</mml:mo><mml:mrow><mml:mo>−</mml:mo><mml:msup><mml:mn>1</mml:mn><mml:mo>′</mml:mo></mml:msup></mml:mrow></mml:msup><mml:mo>,</mml:mo></mml:math></disp-formula>
</p>
<p>where <inline-formula id="ieqn-32"><mml:math id="mml-ieqn-32"><mml:msub><mml:mrow><mml:mi mathvariant="bold">B</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> is a non-symmetric <italic>Q </italic>× <italic>Q</italic> matrix containing the unstandardized cluster-specific regression coefficients between LVs, and <inline-formula id="ieqn-33"><mml:math id="mml-ieqn-33"><mml:msub><mml:mi mathvariant="bold">Ψ</mml:mi><mml:mrow><mml:mi>g</mml:mi><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> is the residual factor covariance matrix. The group-and-cluster-specific nature of <inline-formula id="ieqn-34"><mml:math id="mml-ieqn-34"><mml:msub><mml:mi mathvariant="bold">Ψ</mml:mi><mml:mrow><mml:mi>g</mml:mi><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> ensures that the clustering is driven only by the regression coefficients <inline-formula id="ieqn-35"><mml:math id="mml-ieqn-35"><mml:msub><mml:mrow><mml:mi mathvariant="bold">B</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, and not (also) by the residual factor covariances<xref ref-type="fn" rid="fn3"><sup>3</sup></xref><fn id="fn3"><label>3</label>
<p>To this aim, the residual (co)variances of the endogenous factors are group-and-cluster-specific, whereas the (co)variances of the exogenous factors are group-specific. For more details, please see the original paper by <xref ref-type="bibr" rid="r31">Perez Alonso et al. (2024)</xref>.</p></fn>. The SM for each group-cluster combination <italic>gk</italic> is fitted by minimizing the differences between the model-implied factor covariance matrices <inline-formula id="ieqn-36"><mml:math id="mml-ieqn-36"><mml:msub><mml:mi mathvariant="bold">Φ</mml:mi><mml:mrow><mml:mi>g</mml:mi><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> in Step 2 and the group-specific covariance matrices <inline-formula id="ieqn-37"><mml:math id="mml-ieqn-37"><mml:msubsup><mml:mi mathvariant="bold">Φ</mml:mi><mml:mrow><mml:mi>g</mml:mi></mml:mrow><mml:mrow><mml:mi>s</mml:mi><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:math></inline-formula> from Step 1.</p>
<p>For the mixture clustering, MMG-SEM assumes that the vector of factor scores <inline-formula id="ieqn-38"><mml:math id="mml-ieqn-38"><mml:msub><mml:mi mathvariant="bold-italic">η</mml:mi><mml:mrow><mml:msub><mml:mi>n</mml:mi><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:msub></mml:math></inline-formula> is sampled from a mixture of <italic>K</italic> multivariate normal distributions and that all factor scores of Group <italic>g</italic> — gathered in a matrix <inline-formula id="ieqn-39"><mml:math id="mml-ieqn-39"><mml:msub><mml:mrow><mml:mi mathvariant="bold">H</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> of factor scores — are sampled from the same distribution. Specifically, the formal definition for Group <italic>g</italic> is the following 
	<disp-formula id="eqn-5"><label>5</label><mml:math id="mml-eqn-5" display="block"><mml:mi>f</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mrow><mml:mi mathvariant="bold">H</mml:mi></mml:mrow><mml:mi>g</mml:mi></mml:msub><mml:mo>;</mml:mo><mml:mi>ϑ</mml:mi><mml:mo stretchy="false">)</mml:mo><mml:mo>=</mml:mo><mml:munderover><mml:mo>∑</mml:mo><mml:mrow><mml:mi>k</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mi>K</mml:mi></mml:munderover><mml:msub><mml:mi>π</mml:mi><mml:mi>k</mml:mi></mml:msub><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:mi>g</mml:mi><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mrow><mml:mtext>H</mml:mtext></mml:mrow><mml:mi>g</mml:mi></mml:msub><mml:mo>;</mml:mo><mml:msub><mml:mi>ϑ</mml:mi><mml:mrow><mml:mi>g</mml:mi><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:mo stretchy="false">)</mml:mo><mml:mo>=</mml:mo><mml:munderover><mml:mo>∑</mml:mo><mml:mrow><mml:mi>k</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mi>K</mml:mi></mml:munderover><mml:msub><mml:mi>π</mml:mi><mml:mi>k</mml:mi></mml:msub><mml:munderover><mml:mo>∏</mml:mo><mml:mrow><mml:msub><mml:mi>n</mml:mi><mml:mi>g</mml:mi></mml:msub><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:msub><mml:mi>N</mml:mi><mml:mi>g</mml:mi></mml:msub></mml:mrow></mml:munderover><mml:mi>M</mml:mi><mml:mi>V</mml:mi><mml:mi>N</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi mathvariant="bold-italic">η</mml:mi><mml:mrow><mml:msub><mml:mi>n</mml:mi><mml:mi>g</mml:mi></mml:msub></mml:mrow></mml:msub><mml:mo>;</mml:mo><mml:msub><mml:mi mathvariant="bold-italic">α</mml:mi><mml:mi>g</mml:mi></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mi mathvariant="bold">Φ</mml:mi><mml:mrow><mml:mi>g</mml:mi><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:mo stretchy="false">)</mml:mo><mml:mo>,</mml:mo></mml:math></disp-formula>
</p>
<p>where <italic>f</italic> is the population density function, <italic>ϑ</italic> is the set of population parameters, π<sub><italic>k</italic></sub> is the prior probability of a Group <italic>g</italic> belonging to Cluster <italic>k</italic> (where <inline-formula id="ieqn-40"><mml:math id="mml-ieqn-40"><mml:munderover><mml:mo>∑</mml:mo><mml:mrow><mml:mi>k</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>K</mml:mi></mml:mrow></mml:munderover><mml:msub><mml:mi>π</mml:mi><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:math></inline-formula>), <italic>f</italic><sub><italic>gk</italic></sub> is the density function of the Group <italic>g</italic> in the <italic>k</italic>th cluster, and <italic>ϑ</italic><sub><italic>gk</italic></sub> is its corresponding set of parameters. Specifically, <italic>f</italic><sub><italic>gk</italic></sub> is a multivariate normal distribution where <inline-formula id="ieqn-41"><mml:math id="mml-ieqn-41"><mml:msub><mml:mi mathvariant="bold">Φ</mml:mi><mml:mrow><mml:mi>g</mml:mi><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> and <inline-formula id="ieqn-42"><mml:math id="mml-ieqn-42"><mml:msub><mml:mi mathvariant="bold-italic">α</mml:mi><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> are the factors’ covariance matrix and mean vector, respectively. The covariance matrix is decomposed as indicated in <xref ref-type="disp-formula" rid="eqn-4">Equation 4</xref>, and the factor means <inline-formula id="ieqn-43"><mml:math id="mml-ieqn-43"><mml:msub><mml:mi mathvariant="bold-italic">α</mml:mi><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> are equal to zero due to the centering.</p></sec>
<sec><title>Model Estimation</title>
<p>The unknown parameters <italic>ϑ</italic> of Step 2 (<xref ref-type="disp-formula" rid="eqn-5">Equation 5</xref>) are estimated by means of maximum likelihood estimation using an EM algorithm (for details, see <xref ref-type="bibr" rid="r31">Perez Alonso et al., 2024</xref>). Specifically, the following log-likelihood function is maximized: 
<disp-formula id="eqn-6"><label>6</label><mml:math id="mml-eqn-6" display="block"><mml:msub><mml:mrow><mml:mtext>logL</mml:mtext></mml:mrow><mml:mrow><mml:mi mathvariant="bold-italic">η</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:munderover><mml:mo>∑</mml:mo><mml:mrow><mml:mi>g</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mi>G</mml:mi></mml:munderover><mml:mi>l</mml:mi><mml:mi>o</mml:mi><mml:mi>g</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:munderover><mml:mo>∑</mml:mo><mml:mrow><mml:mi>k</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mi>K</mml:mi></mml:munderover><mml:msub><mml:mi>π</mml:mi><mml:mi>k</mml:mi></mml:msub><mml:msup><mml:mrow><mml:mo>(</mml:mo><mml:mfrac><mml:mn>1</mml:mn><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mn>2</mml:mn><mml:mspace width="thinmathspace"/><mml:mi>π</mml:mi><mml:msup><mml:mo stretchy="false">)</mml:mo><mml:mrow><mml:mi>Q</mml:mi><mml:mrow><mml:mo>/</mml:mo></mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup><mml:mrow><mml:mo stretchy="false">|</mml:mo></mml:mrow><mml:msub><mml:mi mathvariant="bold">Φ</mml:mi><mml:mrow><mml:mi>g</mml:mi><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:msup><mml:mrow><mml:mo stretchy="false">|</mml:mo></mml:mrow><mml:mrow><mml:mn>1</mml:mn><mml:mrow><mml:mo>/</mml:mo></mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup></mml:mrow></mml:mfrac><mml:mi>e</mml:mi><mml:mi>x</mml:mi><mml:mi>p</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mo>−</mml:mo><mml:mfrac><mml:mn>1</mml:mn><mml:mn>2</mml:mn></mml:mfrac><mml:mi>t</mml:mi><mml:mi>r</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:msubsup><mml:mi mathvariant="bold">Φ</mml:mi><mml:mi>g</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mn>1</mml:mn></mml:mrow></mml:msubsup><mml:msubsup><mml:mi mathvariant="bold">Φ</mml:mi><mml:mrow><mml:mi>g</mml:mi><mml:mi>k</mml:mi></mml:mrow><mml:mrow><mml:mo>−</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msubsup><mml:mo stretchy="false">)</mml:mo><mml:mo>)</mml:mo></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mrow><mml:msub><mml:mi>N</mml:mi><mml:mi>g</mml:mi></mml:msub></mml:mrow></mml:msup><mml:mo>)</mml:mo></mml:mrow><mml:mo>,</mml:mo></mml:math></disp-formula>
</p>
<p>where <inline-formula id="ieqn-44"><mml:math id="mml-ieqn-44"><mml:msubsup><mml:mi mathvariant="bold">Φ</mml:mi><mml:mrow><mml:mi>g</mml:mi></mml:mrow><mml:mrow><mml:mi>s</mml:mi><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:math></inline-formula> is Step 1’s factor covariance (<xref ref-type="disp-formula" rid="eqn-3">Equation 3</xref>), and <inline-formula id="ieqn-45"><mml:math id="mml-ieqn-45"><mml:msub><mml:mi mathvariant="bold">Φ</mml:mi><mml:mrow><mml:mi>g</mml:mi><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> is Step 2’s factor covariance (<xref ref-type="disp-formula" rid="eqn-4">Equation 4</xref>).</p>
<p>The log-likelihood function in <xref ref-type="disp-formula" rid="eqn-6">Equation 6</xref> considers only the parameters in the SM. The log-likelihood function for the full MMG-SEM model (i.e., combining Step 1 and Step 2) is defined as 
<disp-formula id="eqn-7"><label>7</label><mml:math id="mml-eqn-7" display="block"><mml:mrow><mml:mtext>logL</mml:mtext></mml:mrow><mml:mo>=</mml:mo><mml:munderover><mml:mo>∑</mml:mo><mml:mrow><mml:mi>g</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mi>G</mml:mi></mml:munderover><mml:mi>l</mml:mi><mml:mi>o</mml:mi><mml:mi>g</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:munderover><mml:mo>∑</mml:mo><mml:mrow><mml:mi>k</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mi>K</mml:mi></mml:munderover><mml:msub><mml:mi>π</mml:mi><mml:mi>k</mml:mi></mml:msub><mml:munderover><mml:mo>∏</mml:mo><mml:mrow><mml:msub><mml:mi>n</mml:mi><mml:mi>g</mml:mi></mml:msub><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:msub><mml:mi>N</mml:mi><mml:mi>g</mml:mi></mml:msub></mml:mrow></mml:munderover><mml:mfrac><mml:mn>1</mml:mn><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mn>2</mml:mn><mml:mspace width="thinmathspace"/><mml:mi>π</mml:mi><mml:msup><mml:mo stretchy="false">)</mml:mo><mml:mrow><mml:mi>J</mml:mi><mml:mrow><mml:mo>/</mml:mo></mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup><mml:mrow><mml:mo stretchy="false">|</mml:mo></mml:mrow><mml:msub><mml:mi mathvariant="bold">Σ</mml:mi><mml:mrow><mml:mi>g</mml:mi><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:msup><mml:mrow><mml:mo stretchy="false">|</mml:mo></mml:mrow><mml:mrow><mml:mn>1</mml:mn><mml:mrow><mml:mo>/</mml:mo></mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup></mml:mrow></mml:mfrac><mml:mi>e</mml:mi><mml:mi>x</mml:mi><mml:mi>p</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mo>−</mml:mo><mml:mfrac><mml:mn>1</mml:mn><mml:mn>2</mml:mn></mml:mfrac><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mrow><mml:mtext>x</mml:mtext></mml:mrow><mml:mrow><mml:msub><mml:mi>n</mml:mi><mml:mi>g</mml:mi></mml:msub></mml:mrow></mml:msub><mml:mo>−</mml:mo><mml:msub><mml:mi mathvariant="bold-italic">μ</mml:mi><mml:mi>g</mml:mi></mml:msub><mml:msup><mml:mo stretchy="false">)</mml:mo><mml:mo>′</mml:mo></mml:msup><mml:msubsup><mml:mi mathvariant="bold">Σ</mml:mi><mml:mrow><mml:mi>g</mml:mi><mml:mi>k</mml:mi></mml:mrow><mml:mrow><mml:mo>−</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msubsup><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mrow><mml:mtext>x</mml:mtext></mml:mrow><mml:mrow><mml:msub><mml:mi>n</mml:mi><mml:mi>g</mml:mi></mml:msub></mml:mrow></mml:msub><mml:mo>−</mml:mo><mml:msub><mml:mi mathvariant="bold-italic">μ</mml:mi><mml:mi>g</mml:mi></mml:msub><mml:mo stretchy="false">)</mml:mo><mml:mo>)</mml:mo></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mo>,</mml:mo></mml:math></disp-formula>
</p>
<p>which can also be expressed in terms of covariance matrices to resemble <xref ref-type="disp-formula" rid="eqn-6">Equation 6</xref> as follows 
<disp-formula id="eqn-8"><label>8</label><mml:math id="mml-eqn-8" display="block"><mml:mrow><mml:mtext>logL</mml:mtext></mml:mrow><mml:mo>=</mml:mo><mml:munderover><mml:mo>∑</mml:mo><mml:mrow><mml:mi>g</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mi>G</mml:mi></mml:munderover><mml:mi>l</mml:mi><mml:mi>o</mml:mi><mml:mi>g</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:munderover><mml:mo>∑</mml:mo><mml:mrow><mml:mi>k</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mi>K</mml:mi></mml:munderover><mml:msub><mml:mi>π</mml:mi><mml:mi>k</mml:mi></mml:msub><mml:msup><mml:mrow><mml:mo>(</mml:mo><mml:mfrac><mml:mn>1</mml:mn><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mn>2</mml:mn><mml:mspace width="thinmathspace"/><mml:mi>π</mml:mi><mml:msup><mml:mo stretchy="false">)</mml:mo><mml:mrow><mml:mi>Q</mml:mi><mml:mrow><mml:mo>/</mml:mo></mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup><mml:mrow><mml:mo stretchy="false">|</mml:mo></mml:mrow><mml:msub><mml:mi mathvariant="bold">Σ</mml:mi><mml:mrow><mml:mi>g</mml:mi><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:msup><mml:mrow><mml:mo stretchy="false">|</mml:mo></mml:mrow><mml:mrow><mml:mn>1</mml:mn><mml:mrow><mml:mo>/</mml:mo></mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup></mml:mrow></mml:mfrac><mml:mi>e</mml:mi><mml:mi>x</mml:mi><mml:mi>p</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mo>−</mml:mo><mml:mfrac><mml:mn>1</mml:mn><mml:mn>2</mml:mn></mml:mfrac><mml:mi>t</mml:mi><mml:mi>r</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mrow><mml:mtext>S</mml:mtext></mml:mrow><mml:mi>g</mml:mi></mml:msub><mml:msubsup><mml:mi mathvariant="bold">Σ</mml:mi><mml:mrow><mml:mi>g</mml:mi><mml:mi>k</mml:mi></mml:mrow><mml:mrow><mml:mo>−</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msubsup><mml:mo stretchy="false">)</mml:mo><mml:mo>)</mml:mo></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mrow><mml:msub><mml:mi>N</mml:mi><mml:mi>g</mml:mi></mml:msub></mml:mrow></mml:msup><mml:mo>)</mml:mo></mml:mrow><mml:mo>,</mml:mo></mml:math></disp-formula>
</p>
<p>where <inline-formula id="ieqn-46"><mml:math id="mml-ieqn-46"><mml:msub><mml:mi mathvariant="bold-italic">μ</mml:mi><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> and <inline-formula id="ieqn-47"><mml:math id="mml-ieqn-47"><mml:msub><mml:mi mathvariant="bold">Σ</mml:mi><mml:mrow><mml:mi>g</mml:mi><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> are the mean vector and model-implied covariance matrix of the observed items, respectively. The <inline-formula id="ieqn-48"><mml:math id="mml-ieqn-48"><mml:msub><mml:mi mathvariant="bold-italic">μ</mml:mi><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> is zero due to the centering, and the <inline-formula id="ieqn-49"><mml:math id="mml-ieqn-49"><mml:msub><mml:mi mathvariant="bold">Σ</mml:mi><mml:mrow><mml:mi>g</mml:mi><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> can be reconstructed by inserting <xref ref-type="disp-formula" rid="eqn-4">Equation 4</xref> into <xref ref-type="disp-formula" rid="eqn-3">Equation 3</xref> as 
<disp-formula id="eqn-9"><label>9</label><mml:math id="mml-eqn-9" display="block"><mml:msub><mml:mi mathvariant="bold">Σ</mml:mi><mml:mrow><mml:mi>g</mml:mi><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mi mathvariant="bold">Λ</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mtext>I</mml:mtext></mml:mrow><mml:mo>−</mml:mo><mml:msub><mml:mrow><mml:mtext>B</mml:mtext></mml:mrow><mml:mi>k</mml:mi></mml:msub><mml:msup><mml:mo stretchy="false">)</mml:mo><mml:mrow><mml:mo>−</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msup><mml:msub><mml:mi mathvariant="bold">Ψ</mml:mi><mml:mrow><mml:mi>g</mml:mi><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mtext>I</mml:mtext></mml:mrow><mml:mo>−</mml:mo><mml:msub><mml:mrow><mml:mtext>B</mml:mtext></mml:mrow><mml:mi>k</mml:mi></mml:msub><mml:msup><mml:mo stretchy="false">)</mml:mo><mml:mrow><mml:mo>−</mml:mo><mml:msup><mml:mn>1</mml:mn><mml:mo>′</mml:mo></mml:msup></mml:mrow></mml:msup><mml:msup><mml:mi mathvariant="bold">Λ</mml:mi><mml:mo>′</mml:mo></mml:msup><mml:mo>+</mml:mo><mml:msub><mml:mi mathvariant="bold">Θ</mml:mi><mml:mi>g</mml:mi></mml:msub><mml:mo>.</mml:mo></mml:math></disp-formula>
</p>
<p>Finally, step-wise estimation as the one described above brings many advantages, such as an intuitive approach (i.e., deal with the measurement model first and then study the structural parameters of interest), robustness against local model misspecifications (<xref ref-type="bibr" rid="r33">Rosseel &amp; Loh, 2022</xref>), and a simplified estimation for complex models (<xref ref-type="bibr" rid="r31">Perez Alonso et al., 2024</xref>). However, as explained by <xref ref-type="bibr" rid="r3">Bakk et al. (2014)</xref>, step-wise estimation also introduces uncertainty in the second step (i.e., additional variance in the estimates), which can lead to biased standard errors if not accounted for. A procedure to correct for biased standard errors is described in detail in <xref ref-type="bibr" rid="r3">Bakk et al. (2014)</xref> and <xref ref-type="bibr" rid="r33">Rosseel and Loh (2022)</xref>. The same procedure is currently implemented in MMG-SEM when computing standard errors (see <xref ref-type="bibr" rid="r30.75">Perez Alonso, 2025</xref> for details).</p></sec></sec>
<sec><title>Model Selection</title>
<p>A common challenge for clustering methods is selecting an appropriate number of clusters. Researchers have tried to solve this problem based on different approaches, such as balancing model fit and complexity (<xref ref-type="bibr" rid="r1">Akaike, 1974</xref>; <xref ref-type="bibr" rid="r2">Akogul &amp; Erisoglu, 2016</xref>; <xref ref-type="bibr" rid="r35">Schwarz, 1978</xref>), considering relative fit improvement (<xref ref-type="bibr" rid="r11">Ceulemans &amp; Kiers, 2006</xref>), cluster separation (<xref ref-type="bibr" rid="r5">Biernacki et al., 2000</xref>), and/or substantive interpretation (<xref ref-type="bibr" rid="r37">van den Bergh et al., 2017</xref>). Given the popularity of clustering, new methods for model selection keep emerging. However, we focus on commonly used approaches for model selection in the context of mixture SEM methods, which are detailed below.</p>
<sec><title>Akaike Information Criterion</title>
<p>The Akaike Information Criterion (AIC; <xref ref-type="bibr" rid="r1">Akaike, 1974</xref>) combines the model fit (i.e., the log-likelihood) with a penalty for model complexity (i.e., number of parameters). It is defined as: 
<disp-formula id="eqn-10"><label>10</label><mml:math id="mml-eqn-10" display="block"><mml:mrow><mml:mtext>AIC</mml:mtext></mml:mrow><mml:mo>=</mml:mo><mml:mo>−</mml:mo><mml:mn>2</mml:mn><mml:mspace width="thinmathspace"/><mml:mrow><mml:mtext>logL</mml:mtext></mml:mrow><mml:mo>+</mml:mo><mml:mn>2</mml:mn><mml:mi>P</mml:mi></mml:math></disp-formula>
</p>
<p>where <italic>P</italic> is the number of parameters. For MMG-SEM, <italic>P</italic> is the sum of the number of mixing proportions (minus one restriction), the number of cluster-specific regressions coefficients, the number of group-specific exogenous factor covariances, the number of group-and-cluster-specific endogenous factor covariances<xref ref-type="fn" rid="fn4"><sup>4</sup></xref><fn id="fn4"><label>4</label>
<p>We only count one set of endogenous covariances for each group, given we assume each group belongs to only one cluster. The endogenous covariances (from the clusters the groups do not belong to) are nuisance parameters.</p></fn>, the number of loadings (minus <italic>Q</italic> fixed loadings due to factor scaling and accounting for (non-)invariant loadings), and the number of group-specific unique variances.</p>
<p>A number of modifications of the AIC have been presented. In this paper, we consider only one such modification: the <inline-formula id="ieqn-50"><mml:math id="mml-ieqn-50"><mml:mrow><mml:mi mathvariant="normal">A</mml:mi><mml:mi mathvariant="normal">I</mml:mi><mml:msub><mml:mi mathvariant="normal">C</mml:mi><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:math></inline-formula> (<xref ref-type="bibr" rid="r7">Bozdogan, 1994</xref>), which was developed specifically for determining the number of clusters in mixture models. It is defined as: 
<disp-formula id="eqn-11"><label>11</label><mml:math id="mml-eqn-11" display="block"><mml:mrow><mml:msub><mml:mrow><mml:mtext>AIC</mml:mtext></mml:mrow><mml:mn>3</mml:mn></mml:msub><mml:mo>=</mml:mo><mml:mo>−</mml:mo><mml:mn>2</mml:mn><mml:mi>log</mml:mi><mml:mtext>L</mml:mtext><mml:mo>+</mml:mo><mml:mn>3</mml:mn><mml:mi>P</mml:mi><mml:mo>.</mml:mo></mml:mrow></mml:math></disp-formula>
</p></sec>
<sec><title>Bayesian Information Criterion</title>
<p>The Bayesian Information Criterion (BIC; <xref ref-type="bibr" rid="r35">Schwarz, 1978</xref>) balances model fit and model complexity as follows: 
<disp-formula id="eqn-12"><label>12</label><mml:math id="mml-eqn-12" display="block"><mml:mrow><mml:mtext>BIC</mml:mtext></mml:mrow><mml:mo>=</mml:mo><mml:mo>−</mml:mo><mml:mn>2</mml:mn><mml:mspace width="thinmathspace"/><mml:mrow><mml:mtext>logL</mml:mtext></mml:mrow><mml:mo>+</mml:mo><mml:mi>P</mml:mi><mml:mrow><mml:mtext>log</mml:mtext></mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mi>S</mml:mi><mml:mi>S</mml:mi><mml:mo stretchy="false">)</mml:mo><mml:mo>,</mml:mo></mml:math></disp-formula>
</p>
<p>where the penalty of the model complexity is now weighted by the logarithm of the sample size <italic>SS</italic>. Usually, the total number of observations <italic>N</italic> is used as the <italic>SS</italic> <inline-formula id="ieqn-51"><mml:math id="mml-ieqn-51"><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mrow><mml:mi mathvariant="normal">B</mml:mi><mml:mi mathvariant="normal">I</mml:mi><mml:mi mathvariant="normal">C</mml:mi></mml:mrow><mml:mrow><mml:mi>N</mml:mi></mml:mrow></mml:msub><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula>, but it has been suggested to use the number of Groups <italic>G</italic> instead of <italic>N</italic> <inline-formula id="ieqn-52"><mml:math id="mml-ieqn-52"><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mrow><mml:mi mathvariant="normal">B</mml:mi><mml:mi mathvariant="normal">I</mml:mi><mml:mi mathvariant="normal">C</mml:mi></mml:mrow><mml:mrow><mml:mi>G</mml:mi></mml:mrow></mml:msub><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula> when selecting the number of group-level clusters (<xref ref-type="bibr" rid="r24">Lukočienė et al., 2010</xref>; <xref ref-type="bibr" rid="r25">Lukočienė &amp; Vermunt, 2009</xref>). <xref ref-type="bibr" rid="r15">De Roover (2021)</xref> and <xref ref-type="bibr" rid="r16">De Roover et al. (2022)</xref> found a superior performance of <inline-formula id="ieqn-53"><mml:math id="mml-ieqn-53"><mml:msub><mml:mrow><mml:mi mathvariant="normal">B</mml:mi><mml:mi mathvariant="normal">I</mml:mi><mml:mi mathvariant="normal">C</mml:mi></mml:mrow><mml:mrow><mml:mi>G</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> in the context of a related mixture multigroup approach.</p></sec>
<sec><title>Convex Hull</title>
<p>The convex hull procedure (CHull; <xref ref-type="bibr" rid="r11">Ceulemans &amp; Kiers, 2006</xref>) has been shown to be a valid alternative to BIC and AIC in the context of mixtures of factor analyzers (<xref ref-type="bibr" rid="r8">Bulteel et al., 2013</xref>). CHull is a generalized scree test that balances model fit and model complexity by plotting the <italic>logL</italic> of the different models in function of their number of parameters <italic>P</italic>. Then, for each model on convex hull of the scree plot, a scree ratio is computed and the solution with the maximal scree ratio is selected. Specifically, the scree ratio <italic>sr</italic><sub><italic>d</italic></sub> for model <italic>d</italic> is defined as: 
<disp-formula id="eqn-13"><label>13</label><mml:math id="mml-eqn-13" display="block"><mml:mi>s</mml:mi><mml:msub><mml:mi>r</mml:mi><mml:mi>d</mml:mi></mml:msub><mml:mo>=</mml:mo><mml:mrow><mml:mo fence="true" stretchy="true" symmetric="true"/><mml:mfrac><mml:mrow><mml:msub><mml:mrow><mml:mtext>logL</mml:mtext></mml:mrow><mml:mi>d</mml:mi></mml:msub><mml:mo>−</mml:mo><mml:msub><mml:mrow><mml:mtext>logL</mml:mtext></mml:mrow><mml:mrow><mml:mi>d</mml:mi><mml:mo>−</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:msub><mml:mi>P</mml:mi><mml:mi>d</mml:mi></mml:msub><mml:mo>−</mml:mo><mml:msub><mml:mi>P</mml:mi><mml:mrow><mml:mi>d</mml:mi><mml:mo>−</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:mfrac><mml:mrow/><mml:mo stretchy="true">/</mml:mo><mml:mrow/><mml:mfrac><mml:mrow><mml:msub><mml:mrow><mml:mtext>logL</mml:mtext></mml:mrow><mml:mrow><mml:mi>d</mml:mi><mml:mo>+</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>−</mml:mo><mml:msub><mml:mrow><mml:mtext>logL</mml:mtext></mml:mrow><mml:mi>d</mml:mi></mml:msub></mml:mrow><mml:mrow><mml:msub><mml:mi>P</mml:mi><mml:mrow><mml:mi>d</mml:mi><mml:mo>+</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>−</mml:mo><mml:msub><mml:mi>P</mml:mi><mml:mi>d</mml:mi></mml:msub></mml:mrow></mml:mfrac><mml:mo fence="true" stretchy="true" symmetric="true"/></mml:mrow><mml:mo>,</mml:mo></mml:math></disp-formula>
</p>
<p>where <inline-formula id="ieqn-54"><mml:math id="mml-ieqn-54"><mml:mi>d</mml:mi><mml:mo>−</mml:mo><mml:mn>1</mml:mn></mml:math></inline-formula> refers to the previous (less complex) model on the hull and <inline-formula id="ieqn-55"><mml:math id="mml-ieqn-55"><mml:mi>d</mml:mi><mml:mo>+</mml:mo><mml:mn>1</mml:mn></mml:math></inline-formula> refers to the next (more complex) model on the hull. It is worth noting that a scree ratio cannot be computed for the least complex model, so it will always select a model with at least two clusters. However, if no clear elbow in the scree plot, one may still conclude that an underlying cluster is unlikely.</p></sec>
<sec><title>Integrated Completed Likelihood</title>
<p>The Integrated Completed Likelihood (ICL; <xref ref-type="bibr" rid="r5">Biernacki et al., 2000</xref>) is a model selection criterion developed for mixture clustering as an alternative to the BIC. The BIC does not consider an essential aspect of the mixture models; that is, the estimated cluster memberships <inline-formula id="ieqn-56"><mml:math id="mml-ieqn-56"><mml:msub><mml:mrow><mml:mover><mml:mi>z</mml:mi><mml:mo stretchy="false">^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>g</mml:mi><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>. Therefore, <xref ref-type="bibr" rid="r5">Biernacki et al. (2000)</xref> proposed using the <italic>Entropy</italic>, which is a measure of the uncertainty of Group <italic>g</italic> belonging to Cluster <italic>k</italic>. Remember that <inline-formula id="ieqn-57"><mml:math id="mml-ieqn-57"><mml:msub><mml:mrow><mml:mover><mml:mi>z</mml:mi><mml:mo stretchy="false">^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>g</mml:mi><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> is a probability ranging from 0 to 1. Formally, for MMG-SEM, the <italic>Entropy</italic> can be defined as: 
<disp-formula id="eqn-14"><label>14</label><mml:math id="mml-eqn-14" display="block"><mml:mi>E</mml:mi><mml:mi>n</mml:mi><mml:mi>t</mml:mi><mml:mi>r</mml:mi><mml:mi>o</mml:mi><mml:mi>p</mml:mi><mml:mi>y</mml:mi><mml:mo>=</mml:mo><mml:munderover><mml:mo>∑</mml:mo><mml:mrow><mml:mi>g</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mi>G</mml:mi></mml:munderover><mml:munderover><mml:mo>∑</mml:mo><mml:mrow><mml:mi>k</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mi>K</mml:mi></mml:munderover><mml:mo stretchy="false">(</mml:mo><mml:mo>−</mml:mo><mml:msub><mml:mrow><mml:mover><mml:mi>z</mml:mi><mml:mo stretchy="false">^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>g</mml:mi><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:mo stretchy="false">)</mml:mo><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mtext>log</mml:mtext></mml:mrow><mml:msub><mml:mrow><mml:mover><mml:mi>z</mml:mi><mml:mo stretchy="false">^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>g</mml:mi><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:mo stretchy="false">)</mml:mo></mml:math></disp-formula>
</p>
<p>The complete derivation of the ICL can be found in <xref ref-type="bibr" rid="r5">Biernacki et al. (2000)</xref>, but its approximation, based on the BIC, is rather simple. Formally, the ICL is approximated as<xref ref-type="fn" rid="fn5"><sup>5</sup></xref><fn id="fn5"><label>5</label>
<p><xref ref-type="bibr" rid="r5">Biernacki et al. (2000)</xref> defined the ICL slightly differently as <inline-formula id="ieqn-162"><mml:math id="mml-ieqn-162"><mml:mrow><mml:mi mathvariant="normal">B</mml:mi><mml:mi mathvariant="normal">I</mml:mi><mml:mi mathvariant="normal">C</mml:mi></mml:mrow><mml:mo>−</mml:mo><mml:mi>E</mml:mi><mml:mi>n</mml:mi><mml:mi>t</mml:mi><mml:mi>r</mml:mi><mml:mi>o</mml:mi><mml:mi>p</mml:mi><mml:mi>y</mml:mi></mml:math></inline-formula>, but they defined <inline-formula id="ieqn-163"><mml:math id="mml-ieqn-163"><mml:mrow><mml:mi mathvariant="normal">B</mml:mi><mml:mi mathvariant="normal">I</mml:mi><mml:mi mathvariant="normal">C</mml:mi></mml:mrow><mml:mo>=</mml:mo><mml:mrow><mml:mi mathvariant="normal">l</mml:mi><mml:mi mathvariant="normal">o</mml:mi><mml:mi mathvariant="normal">g</mml:mi><mml:mi mathvariant="normal">L</mml:mi></mml:mrow><mml:mo>−</mml:mo><mml:mfrac><mml:mi>P</mml:mi><mml:mn>2</mml:mn></mml:mfrac><mml:mrow><mml:mi mathvariant="normal">l</mml:mi><mml:mi mathvariant="normal">o</mml:mi><mml:mi mathvariant="normal">g</mml:mi></mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mi>S</mml:mi><mml:mi>S</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula> instead of <inline-formula id="ieqn-164"><mml:math id="mml-ieqn-164"><mml:mrow><mml:mi mathvariant="normal">B</mml:mi><mml:mi mathvariant="normal">I</mml:mi><mml:mi mathvariant="normal">C</mml:mi></mml:mrow><mml:mo>=</mml:mo><mml:mo>−</mml:mo><mml:mn>2</mml:mn><mml:mspace width="thinmathspace"/><mml:mrow><mml:mi mathvariant="normal">l</mml:mi><mml:mi mathvariant="normal">o</mml:mi><mml:mi mathvariant="normal">g</mml:mi><mml:mi mathvariant="normal">L</mml:mi></mml:mrow><mml:mo>+</mml:mo><mml:mi>P</mml:mi><mml:mrow><mml:mi mathvariant="normal">l</mml:mi><mml:mi mathvariant="normal">o</mml:mi><mml:mi mathvariant="normal">g</mml:mi></mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mi>S</mml:mi><mml:mi>S</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula>. Therefore, both ICL definitions will lead to the same results in terms of model selection.</p></fn>:
<disp-formula id="eqn-15"><label>15</label><mml:math id="mml-eqn-15" display="block"><mml:mrow><mml:mtext>ICL</mml:mtext></mml:mrow><mml:mo>=</mml:mo><mml:mrow><mml:mtext>BIC</mml:mtext></mml:mrow><mml:mo>+</mml:mo><mml:mn>2</mml:mn><mml:mi>E</mml:mi><mml:mi>n</mml:mi><mml:mi>t</mml:mi><mml:mi>r</mml:mi><mml:mi>o</mml:mi><mml:mi>p</mml:mi><mml:mi>y</mml:mi></mml:math></disp-formula>
</p>
<p>For brevity, we focus on the ICL based on <inline-formula id="ieqn-58"><mml:math id="mml-ieqn-58"><mml:msub><mml:mrow><mml:mi mathvariant="normal">B</mml:mi><mml:mi mathvariant="normal">I</mml:mi><mml:mi mathvariant="normal">C</mml:mi></mml:mrow><mml:mrow><mml:mi>G</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> in the Simulation Study, since this has been shown to perform better than <inline-formula id="ieqn-59"><mml:math id="mml-ieqn-59"><mml:msub><mml:mrow><mml:mi mathvariant="normal">B</mml:mi><mml:mi mathvariant="normal">I</mml:mi><mml:mi mathvariant="normal">C</mml:mi></mml:mrow><mml:mrow><mml:mi>N</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> in the context of group-level clustering.</p></sec></sec></sec>
<sec sec-type="Simulation_Study" id="sim_stu"><title>Simulation Study</title>
<sec><title>Design</title>
<p>The aim of the simulation study was to compare the performance of six model selection measures in selecting the number of clusters for MMG-SEM: AIC, <inline-formula id="ieqn-60"><mml:math id="mml-ieqn-60"><mml:msub><mml:mrow><mml:mi mathvariant="normal">A</mml:mi><mml:mi mathvariant="normal">I</mml:mi><mml:mi mathvariant="normal">C</mml:mi></mml:mrow><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msub></mml:math></inline-formula>, <inline-formula id="ieqn-61"><mml:math id="mml-ieqn-61"><mml:msub><mml:mrow><mml:mi mathvariant="normal">B</mml:mi><mml:mi mathvariant="normal">I</mml:mi><mml:mi mathvariant="normal">C</mml:mi></mml:mrow><mml:mrow><mml:mi>G</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, <inline-formula id="ieqn-62"><mml:math id="mml-ieqn-62"><mml:msub><mml:mrow><mml:mi mathvariant="normal">B</mml:mi><mml:mi mathvariant="normal">I</mml:mi><mml:mi mathvariant="normal">C</mml:mi></mml:mrow><mml:mrow><mml:mi>N</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, CHull, and ICL. To this end, we used a Monte-Carlo simulation with six manipulated factors that are expected to affect the model selection performance. The factors and their corresponding levels are described below:

<list list-type="order">
<list-item>
<p>Size of regression parameters β: 0.3, 0.4;</p></list-item>
<list-item>
<p>Number of groups <italic>G</italic>: 24, 48;</p></list-item>
<list-item>
<p>Within-group sample size <italic>N</italic><sub><italic>g</italic></sub>: 50, 100, 200;</p></list-item>
<list-item>
<p>Number of clusters <italic>K</italic>: 2, 4;</p></list-item>
<list-item>
<p>Cluster size: balanced, unbalanced;</p></list-item>
<list-item>
<p>Within-cluster differences <inline-formula id="ieqn-63"><mml:math id="mml-ieqn-63"><mml:msub><mml:mi>σ</mml:mi><mml:mrow><mml:mi>β</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>: no difference (0), small (0.05), large (0.1).</p></list-item>
</list></p>
<p>The size of the regression parameters (i.e., 0.3 and 0.4) was inspired by previous simulation studies in SEM (e.g., <xref ref-type="bibr" rid="r19">Guenole &amp; Brown, 2014</xref>; <xref ref-type="bibr" rid="r31">Perez Alonso et al., 2024</xref>) and cross-national empirical studies with many groups (e.g., <xref ref-type="bibr" rid="r4">Bastian et al., 2014</xref>; <xref ref-type="bibr" rid="r23">Kuppens et al., 2008</xref>). Similarly, the number of groups was 24 and 48, which correspond to a range of groups that is generally found in empirical large-scale international surveys (<xref ref-type="bibr" rid="r34">Rutkowski &amp; Svetina, 2014</xref>). The number of clusters (i.e., 2 and 4)<xref ref-type="fn" rid="fn6"><sup>6</sup></xref><fn id="fn6"><label>6</label>
<p>A smaller simulation where <italic>K</italic> = 1 was also performed but not included in the main text due to space constraints. More details can be found in <xref ref-type="bibr" rid="r31.5">Perez Alonso et al. (2025)</xref></p></fn> and cluster size were defined considering previous simulation studies on model selection (<xref ref-type="bibr" rid="r5">Biernacki et al., 2000</xref>; <xref ref-type="bibr" rid="r8">Bulteel et al., 2013</xref>; <xref ref-type="bibr" rid="r15">De Roover, 2021</xref>; <xref ref-type="bibr" rid="r16">De Roover et al., 2022</xref>; <xref ref-type="bibr" rid="r24">Lukočienė et al., 2010</xref>; <xref ref-type="bibr" rid="r31">Perez Alonso et al., 2024</xref>; <xref ref-type="bibr" rid="r36">Steinley &amp; Brusco, 2011</xref>) as we want to relate our results to theirs.</p>
<p>In total, the design included 2 (size of regression parameters) <italic>×</italic> 2 (number of groups) <italic>×</italic> 3 (within-group sample size) <italic>×</italic> 2 (number of clusters) <italic>×</italic> 2 (cluster size) <italic>×</italic> 3 (within-cluster differences) = 144 data generation conditions. For all conditions, 100 different data sets were generated, for a total of 14400 data sets. To evaluate the model selection measures, each data set was analyzed six times with MMG-SEM from one to six clusters, for a total of 14400 <italic>×</italic> 6 = 86400 analyses. Note that we added non-invariances to the loadings (see the <xref ref-type="sec" rid="Data_Generation">Data Generation</xref> section for more information) and that such non-invariances are correctly modeled in MMG-SEM. All the data generation and analyses were done using R Version 4.3.3 (<xref ref-type="bibr" rid="r32">R Core Team, 2024</xref>), and the code is openly shared at <xref ref-type="bibr" rid="r30.5">Perez Alonso (2024)</xref>. The data generation procedure is described below.</p></sec>
<sec id="Data_Generation"><title>Data Generation</title>
	<p>Each data set was generated according to the SEM model in <xref ref-type="fig" rid="fig1">Figure 1</xref>, with four LVs, each one measured by five indicators, for a total of 20 observed variables. The structural relations, which are the parameters of interest for the clustering, are represented by four regression parameters. F1 and F2 served as exogenous variables, while F3 and F4 were endogenous variables. Note that F3 acts as a ‘mediator’ and is, thus, an exogenous and endogenous variable at the same time.</p><fig id="fig1" position="anchor" orientation="portrait"><label>Figure 1</label><caption><title>The Model Used for the Data Generation</title><p><italic>Note</italic>. F1 and F2 are exogenous variables, F3 is dependent and independent at the same time (‘mediator’), and F4 is a dependent only variable.</p></caption><graphic mimetype="image" mime-subtype="png" xlink:href="meth.14931-f1.png" position="anchor" orientation="portrait"/></fig>
<p>The sample size per Group <italic>N</italic><sub><italic>g</italic></sub>, the number of Groups <italic>G</italic>, and the number of clusters <italic>K</italic> were defined according to the manipulated factors ‘within-group sample size’, ‘number of groups’, and ‘number of clusters’, respectively. The number of groups per cluster was manipulated according to the ‘cluster size’. In the balanced condition, the groups were equally divided per cluster. For instance, for <italic>G</italic> = 48 and <italic>K</italic> = 4, each cluster contained 12 groups. In contrast, in the unbalanced condition, there was one larger cluster with 75% of the groups, and the remaining clusters were equally sized. For example, for <italic>G</italic> = 48 and <italic>K</italic> = 4, the first cluster would contain 36 groups and the remaining three clusters would contain four groups each.</p>
<p>The 20 observed variables were generated from a <inline-formula id="ieqn-64"><mml:math id="mml-ieqn-64"><mml:mi>M</mml:mi><mml:mi>V</mml:mi><mml:mi>N</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi mathvariant="bold-italic">μ</mml:mi><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mi mathvariant="bold">Σ</mml:mi><mml:mrow><mml:mi>g</mml:mi><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula>, where the mean vector <inline-formula id="ieqn-65"><mml:math id="mml-ieqn-65"><mml:msub><mml:mi mathvariant="bold-italic">μ</mml:mi><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> was a vector of zeros and the covariance matrices <inline-formula id="ieqn-66"><mml:math id="mml-ieqn-66"><mml:msub><mml:mi mathvariant="bold">Σ</mml:mi><mml:mrow><mml:mi>g</mml:mi><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> were generated according to <xref ref-type="disp-formula" rid="eqn-9">Equation 9</xref>. Thus, to generate the data, the parameters in <xref ref-type="disp-formula" rid="eqn-9">Equation 9</xref> (i.e., <inline-formula id="ieqn-67"><mml:math id="mml-ieqn-67"><mml:mi mathvariant="bold">Λ</mml:mi><mml:mo>,</mml:mo><mml:msub><mml:mi mathvariant="bold">Θ</mml:mi><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mrow><mml:mi mathvariant="bold">B</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, and <inline-formula id="ieqn-68"><mml:math id="mml-ieqn-68"><mml:msub><mml:mi mathvariant="bold">Ψ</mml:mi><mml:mrow><mml:mi>g</mml:mi><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>) must be defined. The <inline-formula id="ieqn-69"><mml:math id="mml-ieqn-69"><mml:mi mathvariant="bold">Λ</mml:mi></mml:math></inline-formula> and <inline-formula id="ieqn-70"><mml:math id="mml-ieqn-70"><mml:msub><mml:mi mathvariant="bold">Θ</mml:mi><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> matrices were generated aiming to obtain a total variance per item around 1 and a reliability (<italic>R</italic><sup>2</sup>) of 0.6 for each observed indicator. To do this, the non-zero values in <inline-formula id="ieqn-71"><mml:math id="mml-ieqn-71"><mml:mi mathvariant="bold">Λ</mml:mi></mml:math></inline-formula> were set to <inline-formula id="ieqn-72"><mml:math id="mml-ieqn-72"><mml:msqrt><mml:mn>0.6</mml:mn></mml:msqrt></mml:math></inline-formula> while the residual variances in <inline-formula id="ieqn-73"><mml:math id="mml-ieqn-73"><mml:msub><mml:mi mathvariant="bold">Θ</mml:mi><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> were drawn from a uniform distribution <italic>U</italic>(0.3, 0.5). Note that the residual variances in <inline-formula id="ieqn-74"><mml:math id="mml-ieqn-74"><mml:msub><mml:mi mathvariant="bold">Θ</mml:mi><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> were sampled for each group, allowing group-specific differences as specified in <xref ref-type="disp-formula" rid="eqn-9">Equation 9</xref>.</p>
<p>To evaluate the effect of loading non-invariances on the model selection, we also added between-group differences to the <inline-formula id="ieqn-75"><mml:math id="mml-ieqn-75"><mml:mi mathvariant="bold">Λ</mml:mi></mml:math></inline-formula> matrices. In particular, 50% of the groups presented non-invariances. For each non-invariant group, we applied the non-invariance to the second and third loading of each factor (the first loading is fixed to 1). The non-invariances were randomly sampled from a uniform distribution around 0.4, i.e., <italic>U</italic>(0.3, 0.5), and it was randomly decided whether the non-invariance was added or subtracted to the original loading (i.e., <inline-formula id="ieqn-76"><mml:math id="mml-ieqn-76"><mml:msqrt><mml:mn>0.6</mml:mn></mml:msqrt></mml:math></inline-formula>). As a result, each non-invariant loading was different for each non-invariant group.</p>
<p>The setup of the regression parameters in <inline-formula id="ieqn-77"><mml:math id="mml-ieqn-77"><mml:msub><mml:mrow><mml:mi mathvariant="bold">B</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> can be seen in <xref ref-type="fig" rid="fig2">Figure 2</xref>. The manipulated factor ‘size of the regression parameters’ (β) indicated the size of the coefficients β<sub>1</sub>, β<sub>2</sub>, β<sub>3</sub>, and β<sub>4</sub>. The difference between the clusters was created by setting one of those coefficients to zero in each cluster. Thus, the size of the regression parameters also defined the size of the difference between the clusters. Note that, when <italic>K</italic> = 2, Models 3 and 4 in <xref ref-type="fig" rid="fig2">Figure 2</xref> were not applicable. The parameters in <inline-formula id="ieqn-78"><mml:math id="mml-ieqn-78"><mml:msub><mml:mrow><mml:mi mathvariant="bold">B</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> were also affected by the manipulated factor ‘within-cluster differences’ <inline-formula id="ieqn-79"><mml:math id="mml-ieqn-79"><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi>σ</mml:mi><mml:mrow><mml:mi>β</mml:mi></mml:mrow></mml:msub><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula>. To simulate empirically realistic conditions, we added small differences in the coefficients β to each Group <italic>g</italic> within a Cluster <italic>k</italic>. To do this, within each cluster, the regression parameter of each group was drawn from a normal distribution <inline-formula id="ieqn-80"><mml:math id="mml-ieqn-80"><mml:mi>N</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mi>β</mml:mi><mml:mo>,</mml:mo><mml:msub><mml:mi>σ</mml:mi><mml:mrow><mml:mi>β</mml:mi></mml:mrow></mml:msub><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula>, where the β and the <inline-formula id="ieqn-81"><mml:math id="mml-ieqn-81"><mml:msub><mml:mi>σ</mml:mi><mml:mrow><mml:mi>β</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> acted as the mean and the standard deviation of the distribution, respectively. The values of <inline-formula id="ieqn-82"><mml:math id="mml-ieqn-82"><mml:msub><mml:mi>σ</mml:mi><mml:mrow><mml:mi>β</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> implied no within-cluster differences <inline-formula id="ieqn-83"><mml:math id="mml-ieqn-83"><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi>σ</mml:mi><mml:mrow><mml:mi>β</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mn>0</mml:mn><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula>, small differences <inline-formula id="ieqn-84"><mml:math id="mml-ieqn-84"><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi>σ</mml:mi><mml:mrow><mml:mi>β</mml:mi></mml:mrow></mml:msub> <mml:mo>=</mml:mo> <mml:mn>0.05</mml:mn><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula>, or large differences <inline-formula id="ieqn-85"><mml:math id="mml-ieqn-85"><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi>σ</mml:mi><mml:mrow><mml:mi>β</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mn>0.1</mml:mn><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula>. Note that we expect 99% of the sampled values to be within three standard deviations of the mean when drawing from a normal distribution. For instance, if we consider β = 0.3 and <inline-formula id="ieqn-86"><mml:math id="mml-ieqn-86"><mml:msub><mml:mi>σ</mml:mi><mml:mrow><mml:mi>β</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mn>0.05</mml:mn></mml:math></inline-formula> and we focus on β<sub>1</sub> (see <xref ref-type="fig" rid="fig2">Figure 2</xref>), the values of β<sub>1</sub> in Cluster 1 will be sampled from <italic>N</italic>(0, 0.05), whereas they will be sampled from <italic>N</italic>(0.3, 0.05) in Cluster 2. We expect 99% of the values to lie between -0.15 and 0.15 in Cluster 1 and between 0.15 and 0.45 in Cluster 2. Thus, the small differences level <inline-formula id="ieqn-87"><mml:math id="mml-ieqn-87"><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi>σ</mml:mi><mml:mrow><mml:mi>β</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mn>0.05</mml:mn><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula> led to almost no overlap between clusters, whereas the large level <inline-formula id="ieqn-88"><mml:math id="mml-ieqn-88"><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi>σ</mml:mi><mml:mrow><mml:mi>β</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mn>0.1</mml:mn><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula> ensured overlap between the clusters.</p><fig id="fig2" position="anchor" orientation="portrait"><label>Figure 2</label><caption><title>Zero and Non-Zero Regression Parameters Between the LVs Depending On the Cluster</title></caption><graphic mimetype="image" mime-subtype="png" xlink:href="meth.14931-f2.png" position="anchor" orientation="portrait"/></fig>
<p>Finally, the elements in <inline-formula id="ieqn-89"><mml:math id="mml-ieqn-89"><mml:msub><mml:mi mathvariant="bold">Ψ</mml:mi><mml:mrow><mml:mi>g</mml:mi><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> were defined by sampling the variance of the exogenous variables F1 and F2 from a uniform distribution <italic>U</italic>(0.75, 1.25), and their covariance from <inline-formula id="ieqn-90"><mml:math id="mml-ieqn-90"><mml:mi>U</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mo>−</mml:mo><mml:mn>0.3</mml:mn><mml:mo>,</mml:mo><mml:mn>0.3</mml:mn><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula> for all groups. Similarly, the total variance of the endogenous factors F3 and F4 was sampled from <italic>U</italic>(0.75, 1.25) for each group, and their residual variance depended on the cluster-specific regression parameters. For example, if the total variance of F3 for Group <italic>g</italic> was <italic>Var</italic><sub><italic>tot</italic></sub>, the residual variance <italic>Var</italic><sub><italic>res</italic></sub> for Group <italic>g</italic> and Cluster <italic>k</italic> was <inline-formula id="ieqn-91"><mml:math id="mml-ieqn-91"><mml:mi>V</mml:mi><mml:mi>a</mml:mi><mml:msub><mml:mi>r</mml:mi><mml:mrow><mml:mi>r</mml:mi><mml:mi>e</mml:mi><mml:mi>s</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mi>V</mml:mi><mml:mi>a</mml:mi><mml:msub><mml:mi>r</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>o</mml:mi><mml:mi>t</mml:mi></mml:mrow></mml:msub><mml:mo>−</mml:mo><mml:mo stretchy="false">(</mml:mo><mml:msubsup><mml:mi>β</mml:mi><mml:mrow><mml:mn>2</mml:mn></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msubsup><mml:mspace width="thinmathspace"/><mml:mrow><mml:mi mathvariant="normal">V</mml:mi><mml:mi mathvariant="normal">a</mml:mi><mml:mi mathvariant="normal">r</mml:mi></mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mi>F</mml:mi><mml:mn>1</mml:mn><mml:mo stretchy="false">)</mml:mo><mml:mo>+</mml:mo><mml:msubsup><mml:mi>β</mml:mi><mml:mrow><mml:mn>3</mml:mn></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msubsup><mml:mspace width="thinmathspace"/><mml:mrow><mml:mi mathvariant="normal">V</mml:mi><mml:mi mathvariant="normal">a</mml:mi><mml:mi mathvariant="normal">r</mml:mi></mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mi>F</mml:mi><mml:mn>2</mml:mn><mml:mo stretchy="false">)</mml:mo><mml:mo>+</mml:mo><mml:mn>2</mml:mn><mml:mspace width="thinmathspace"/><mml:msub><mml:mi>β</mml:mi><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub><mml:mspace width="thinmathspace"/><mml:msub><mml:mi>β</mml:mi><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msub><mml:mspace width="thinmathspace"/><mml:mrow><mml:mi mathvariant="normal">C</mml:mi><mml:mi mathvariant="normal">o</mml:mi><mml:mi mathvariant="normal">v</mml:mi></mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mi>F</mml:mi><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mi>F</mml:mi><mml:mn>2</mml:mn><mml:mo stretchy="false">)</mml:mo><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula>.</p></sec>
<sec id="s3_3"><title>R<sup>2</sup> Entropy</title>
<p>To determine the cluster separation in the simulated datasets, we applied the <italic>R</italic><sup>2</sup> Entropy, an <italic>Entropy</italic>-based measure. Specifically, the <italic>R</italic><sup>2</sup> Entropy indicates how well the observed responses predict the cluster memberships <inline-formula id="ieqn-92"><mml:math id="mml-ieqn-92"><mml:msub><mml:mrow><mml:mover><mml:mi>z</mml:mi><mml:mo stretchy="false">^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>g</mml:mi><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> (for details on its calculation, see <xref ref-type="bibr" rid="r40">Vermunt &amp; Magidson, 2016</xref>). It takes on a value of 1 when the clusters are perfectly separated (i.e., no classification uncertainty) and a value of 0 when there is no separation at all.</p>
<p>To get an overview of how the cluster separation was affected by the simulation conditions, we evaluated the <italic>R</italic><sup>2</sup> Entropy at (an approximated) population level. For this, we generated data for each simulation condition with a large sample size. Given that MMG-SEM’s clustering is at the group-level, the relevant sample size for the clustering and <italic>R</italic><sup>2</sup> Entropy is the number of groups <italic>G</italic>, which was set to 192 groups. Subsequently, we computed the cluster memberships <inline-formula id="ieqn-93"><mml:math id="mml-ieqn-93"><mml:msub><mml:mrow><mml:mover><mml:mi>z</mml:mi><mml:mo stretchy="false">^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>g</mml:mi><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> based on the true parameters values for <inline-formula id="ieqn-94"><mml:math id="mml-ieqn-94"><mml:mi mathvariant="bold">Λ</mml:mi><mml:mo>,</mml:mo><mml:msub><mml:mi mathvariant="bold">Θ</mml:mi><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mrow><mml:mi mathvariant="bold">B</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, and <inline-formula id="ieqn-95"><mml:math id="mml-ieqn-95"><mml:msub><mml:mi mathvariant="bold">Ψ</mml:mi><mml:mrow><mml:mi>g</mml:mi><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> (i.e., they were used as starting values in an MMG-SEM analysis where no parameter updates were performed). Per condition, 300 replications were generated.</p>
<p>The <italic>R</italic><sup>2</sup> Entropy ranged from 0.76 to 1 with an average of 0.93 across all data sets, which indicated well-separated clusters overall. The <italic>R</italic><sup>2</sup> Entropy was mostly influenced by the within-cluster differences <inline-formula id="ieqn-96"><mml:math id="mml-ieqn-96"><mml:msub><mml:mi>σ</mml:mi><mml:mrow><mml:mi>β</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, the regression coefficients β, and the within-group sample size <italic>N</italic><sub><italic>g</italic></sub>. Their main effects can be seen in <xref ref-type="table" rid="t1">Table 1</xref>, and the interaction between <inline-formula id="ieqn-97"><mml:math id="mml-ieqn-97"><mml:msub><mml:mi>σ</mml:mi><mml:mrow><mml:mi>β</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> and β is shown in <xref ref-type="fig" rid="fig3">Figure 3</xref>. Lower <italic>R</italic><sup>2</sup> Entropy values were found in difficult conditions involving large within-cluster differences <inline-formula id="ieqn-98"><mml:math id="mml-ieqn-98"><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi>σ</mml:mi><mml:mrow><mml:mi>β</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mn>0.1</mml:mn><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula>, lower regression coefficients (β = 0.3) and/or low within-group sample size <inline-formula id="ieqn-99"><mml:math id="mml-ieqn-99"><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi>N</mml:mi><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mn>50</mml:mn><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula>. It is expected that the model selection measures will struggle to choose the correct number of clusters in conditions where the clusters are not well separated.</p>
<table-wrap id="t1" position="anchor" orientation="portrait">
<label>Table 1</label><caption><title>Average R<sup>2</sup> Entropy per Level of the Most Influential Factors</title></caption>
<table frame="hsides" rules="groups"><colgroup span="1">
	<col width="" align="left"/>
	<col width="" align="left"/>
	<col width="" align="left"/>
	<col width="" align="left"/>
	<col width="" align="left"/>
	<col width="" align="left"/>
	<col width="" align="left"/>
	<col width="" align="left"/></colgroup>
<thead>
<tr>
	<th colspan="3" align="center" valign="baseline"><italic>N</italic><sub><italic>g</italic></sub><hr/></th>
	<th colspan="3" align="center" valign="baseline">σ<sub>β</sub><hr/></th>
	<th colspan="2" align="center" valign="baseline">β<hr/></th>
</tr>
<tr>
<th align="center" >50</th>
<th align="center" >100</th>
<th align="center" >200</th>
<th align="center" >0</th>
<th align="center" >0.05</th>
<th align="center" >0.1</th>
<th align="center" >0.3</th>
<th align="center" >0.4</th>
</tr>
</thead>
<tbody>
<tr>
<td align="char" char=".">0.923</td>
<td align="char" char=".">0.982</td>
<td align="char" char=".">0.993</td>
<td align="char" char=".">0.983</td>
<td align="char" char=".">0.976</td>
<td align="char" char=".">0.938</td>
<td align="char" char=".">0.942</td>
<td align="char" char=".">0.990</td>
</tr>
</tbody>
</table>
</table-wrap><fig id="fig3" position="anchor" orientation="portrait"><label>Figure 3</label><caption><title>Approximated Population <italic>R</italic><sup>2</sup> Entropy in Function of the Within-Cluster Differences <inline-formula id="ieqn-100"><mml:math id="mml-ieqn-100"><mml:msub><mml:mi>σ</mml:mi><mml:mrow><mml:mi>β</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> and the Size of the Regression Parameters β</title></caption><graphic mimetype="image" mime-subtype="png" xlink:href="meth.14931-f3.png" position="anchor" orientation="portrait"/></fig></sec></sec>
<sec sec-type="Results"><title>Results</title>
<p>Before discussing the findings, it is worth noting that the evaluated model selection methods all use the log-likelihood as the measure of model fit. For MMG-SEM, we could use the log-likelihood based on the factors (<xref ref-type="disp-formula" rid="eqn-6">Equation 6</xref>) or the observed data log-likelihood (<xref ref-type="disp-formula" rid="eqn-7">Equation 7</xref>). We evaluated the measures’ performance using both log-likelihoods in our simulation and found only minor differences in the results. Thus, for brevity, we present only the results using the log-likelihood in <xref ref-type="disp-formula" rid="eqn-7">Equation 7</xref>, since this considers how the full model (i.e., measurement + structural model) fits the data. The results using <xref ref-type="disp-formula" rid="eqn-6">Equation 6</xref> can be found in <xref ref-type="bibr" rid="r31.5">Perez Alonso et al. (2025)</xref>.</p>
<sec><title>Model Selection</title>
<p>For each model selection measure, we assessed how often it correctly selected the true number of clusters. To gain a deeper understanding of each measure, we also inspected how often it over- and under-selected the number of clusters. The main effects of the manipulated factors of the simulation can be seen in <xref ref-type="table" rid="t2">Table 2</xref>. In total, the best-performing method was the CHull, followed by the AIC, <inline-formula id="ieqn-102"><mml:math id="mml-ieqn-102"><mml:msub><mml:mrow><mml:mi mathvariant="normal">A</mml:mi><mml:mi mathvariant="normal">I</mml:mi><mml:mi mathvariant="normal">C</mml:mi></mml:mrow><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msub></mml:math></inline-formula>, <inline-formula id="ieqn-103"><mml:math id="mml-ieqn-103"><mml:msub><mml:mrow><mml:mi mathvariant="normal">B</mml:mi><mml:mi mathvariant="normal">I</mml:mi><mml:mi mathvariant="normal">C</mml:mi></mml:mrow><mml:mrow><mml:mi>G</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, ICL, and <inline-formula id="ieqn-104"><mml:math id="mml-ieqn-104"><mml:msub><mml:mrow><mml:mi mathvariant="normal">B</mml:mi><mml:mi mathvariant="normal">I</mml:mi><mml:mi mathvariant="normal">C</mml:mi></mml:mrow><mml:mrow><mml:mi>N</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, with a proportion of correctly selected models of 0.77, 0.66, 0.64, 0.63, 0.61, and 0.54, respectively. Note that the similarity between the results of <inline-formula id="ieqn-105"><mml:math id="mml-ieqn-105"><mml:msub><mml:mrow><mml:mi mathvariant="normal">A</mml:mi><mml:mi mathvariant="normal">I</mml:mi><mml:mi mathvariant="normal">C</mml:mi></mml:mrow><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msub></mml:math></inline-formula> and <inline-formula id="ieqn-106"><mml:math id="mml-ieqn-106"><mml:msub><mml:mrow><mml:mi mathvariant="normal">B</mml:mi><mml:mi mathvariant="normal">I</mml:mi><mml:mi mathvariant="normal">C</mml:mi></mml:mrow><mml:mrow><mml:mi>G</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> can be partially explained by the similarity of their penalties (see <xref ref-type="disp-formula" rid="eqn-11">Equations 11</xref> and <xref ref-type="disp-formula" rid="eqn-12">12</xref>); that is, <inline-formula id="ieqn-107"><mml:math id="mml-ieqn-107"><mml:msub><mml:mrow><mml:mi mathvariant="normal">A</mml:mi><mml:mi mathvariant="normal">I</mml:mi><mml:mi mathvariant="normal">C</mml:mi></mml:mrow><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msub></mml:math></inline-formula>’s penalty is 3, whereas <inline-formula id="ieqn-108"><mml:math id="mml-ieqn-108"><mml:msub><mml:mrow><mml:mi mathvariant="normal">B</mml:mi><mml:mi mathvariant="normal">I</mml:mi><mml:mi mathvariant="normal">C</mml:mi></mml:mrow><mml:mrow><mml:mi>G</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>’s penalty is 3.18 and 3.87 when <italic>G</italic> is 24 and 48, respectively.</p>
<table-wrap id="t2" position="anchor" orientation="portrait">
<label>Table 2</label><caption><title>Proportion of Under-, Over-, and Correct Selection of the Number of Clusters for All Model Selection Measures Per Level of Each Manipulated Factor </title></caption>
<table frame="hsides" rules="groups" style="compact-1"><colgroup span="1">
	<col width="" align="left"/>
	<col width="" align="left"/>
	<col width="" align="left"/>
	<col width="" align="left"/>
	<col width="" align="left"/>
	<col width="" align="left"/>
	<col width="" align="left"/>
	<col width="" align="left"/>
	<col width="" align="left"/>
	<col width="" align="left"/>
	<col width="" align="left"/>
	<col width="" align="left"/>
	<col width="" align="left"/>
	<col width="" align="left"/>
	<col width="" align="left"/>
	<col width="" align="left"/>
	<col width="" align="left"/></colgroup>
<thead>
<tr>
<th/>
	<th/>
<th colspan="2" align="center"><italic>K</italic><hr/></th>
	<th colspan="3" align="center"><italic>N</italic><sub><italic>g</italic></sub><hr/></th>
	<th colspan="2" align="center"><italic>G</italic><hr/></th>
	<th colspan="2" align="center">β<hr/></th>
	<th colspan="2" align="center">Cluster size<hr/></th>
	<th colspan="3" align="center">σ<sub>β</sub><hr/></th>
<th/>
</tr>
<tr>
<th valign="bottom">Measure</th>
<th valign="bottom">Result</th>	
	<th align="center">2</th>
	<th align="center">4</th>
	<th align="center">50</th>
	<th align="center">100</th>
	<th align="center">200</th>
	<th align="center">24</th>
	<th align="center">48</th>
	<th align="center">0.3</th>
	<th align="center">0.4</th>
	<th align="center">Bal</th>
	<th align="center">Unb</th>
	<th align="center">0</th>
	<th align="center">0.05</th>
	<th align="center">0.1</th>
<th valign="bottom">Total</th>	
</tr>
</thead>
<tbody>
<tr>
<td>AIC</td>
<td>Under</td>
<td align="char" char=".">0.02</td>
<td align="char" char=".">0.34</td>
<td align="char" char=".">0.44</td>
<td align="char" char=".">0.10</td>
<td align="char" char=".">0.00</td>
<td align="char" char=".">0.22</td>
<td align="char" char=".">0.14</td>
<td align="char" char=".">0.25</td>
<td align="char" char=".">0.11</td>
<td align="char" char=".">0.12</td>
<td align="char" char=".">0.25</td>
<td align="char" char=".">0.19</td>
<td align="char" char=".">0.19</td>
<td align="char" char=".">0.17</td>
<td align="char" char=".">0.18</td>
</tr>
<tr>
<td/>	
<td>Correct</td>
<td align="char" char=".">0.82</td>
<td align="char" char=".">0.50</td>
<td align="char" char=".">0.56</td>
<td align="char" char=".">0.76</td>
<td align="char" char=".">0.66</td>
<td align="char" char=".">0.64</td>
<td align="char" char=".">0.68</td>
<td align="char" char=".">0.61</td>
<td align="char" char=".">0.71</td>
<td align="char" char=".">0.73</td>
<td align="char" char=".">0.59</td>
<td align="char" char=".">0.80<sup>a</sup></td>
<td align="char" char=".">0.81</td>
<td align="char" char=".">0.37</td>
<td align="char" char=".">0.66</td>
</tr>
	<tr>
	<td/>		
<td>Over</td>
<td align="char" char=".">0.16</td>
<td align="char" char=".">0.16</td>
<td align="char" char=".">0.00</td>
<td align="char" char=".">0.13</td>
<td align="char" char=".">0.34</td>
<td align="char" char=".">0.14</td>
<td align="char" char=".">0.18</td>
<td align="char" char=".">0.14</td>
<td align="char" char=".">0.18</td>
<td align="char" char=".">0.16</td>
<td align="char" char=".">0.16</td>
<td align="char" char=".">0.01</td>
<td align="char" char=".">0.01</td>
<td align="char" char=".">0.46</td>
<td align="char" char=".">0.16</td>
</tr>
<tr style="grey-border-top">
<td><inline-formula id="ieqn-110"><mml:math id="mml-ieqn-110"><mml:msub><mml:mrow><mml:mi mathvariant="normal">A</mml:mi><mml:mi mathvariant="normal">I</mml:mi><mml:mi mathvariant="normal">C</mml:mi></mml:mrow><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msub></mml:math></inline-formula></td>
<td>Under</td>
<td align="char" char=".">0.05</td>
<td align="char" char=".">0.41</td>
<td align="char" char=".">0.52</td>
<td align="char" char=".">0.17</td>
<td align="char" char=".">0.01</td>
<td align="char" char=".">0.30</td>
<td align="char" char=".">0.17</td>
<td align="char" char=".">0.32</td>
<td align="char" char=".">0.15</td>
<td align="char" char=".">0.17</td>
<td align="char" char=".">0.30</td>
<td align="char" char=".">0.24</td>
<td align="char" char=".">0.24</td>
<td align="char" char=".">0.22</td>
<td align="char" char=".">0.23</td>
</tr>
<tr>
<td/>	
<td>Correct</td>
<td align="char" char=".">0.82</td>
<td align="char" char=".">0.46</td>
<td align="char" char=".">0.48</td>
<td align="char" char=".">0.78</td>
<td align="char" char=".">0.67</td>
<td align="char" char=".">0.60</td>
<td align="char" char=".">0.69</td>
<td align="char" char=".">0.57</td>
<td align="char" char=".">0.71</td>
<td align="char" char=".">0.71</td>
<td align="char" char=".">0.57</td>
<td align="char" char=".">0.75</td>
<td align="char" char=".">0.76</td>
<td align="char" char=".">0.42</td>
<td align="char" char=".">0.64</td>
</tr>
<tr>
<td/>
<td>Over</td>
<td align="char" char=".">0.12</td>
<td align="char" char=".">0.12</td>
<td align="char" char=".">0.00</td>
<td align="char" char=".">0.05</td>
<td align="char" char=".">0.32</td>
<td align="char" char=".">0.10</td>
<td align="char" char=".">0.14</td>
<td align="char" char=".">0.11</td>
<td align="char" char=".">0.14</td>
<td align="char" char=".">0.12</td>
<td align="char" char=".">0.13</td>
<td align="char" char=".">0.01</td>
<td align="char" char=".">0.00</td>
<td align="char" char=".">0.36</td>
<td align="char" char=".">0.12</td>
</tr>
	<tr style="grey-border-top">
<td><inline-formula id="ieqn-111"><mml:math id="mml-ieqn-111"><mml:msub><mml:mrow><mml:mi mathvariant="normal">B</mml:mi><mml:mi mathvariant="normal">I</mml:mi><mml:mi mathvariant="normal">C</mml:mi></mml:mrow><mml:mrow><mml:mi>G</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula></td>
<td>Under</td>
<td align="char" char=".">0.07</td>
<td align="char" char=".">0.45</td>
<td align="char" char=".">0.57</td>
<td align="char" char=".">0.20</td>
<td align="char" char=".">0.01</td>
<td align="char" char=".">0.31</td>
<td align="char" char=".">0.21</td>
<td align="char" char=".">0.35</td>
<td align="char" char=".">0.17</td>
<td align="char" char=".">0.19</td>
<td align="char" char=".">0.33</td>
<td align="char" char=".">0.28</td>
<td align="char" char=".">0.27</td>
<td align="char" char=".">0.24</td>
<td align="char" char=".">0.26</td>
</tr>
<tr>
	<td/>
<td>Correct</td>
<td align="char" char=".">0.82</td>
<td align="char" char=".">0.44</td>
<td align="char" char=".">0.43</td>
<td align="char" char=".">0.77</td>
<td align="char" char=".">0.69</td>
<td align="char" char=".">0.59</td>
<td align="char" char=".">0.66</td>
<td align="char" char=".">0.55</td>
<td align="char" char=".">0.70</td>
<td align="char" char=".">0.70</td>
<td align="char" char=".">0.56</td>
<td align="char" char=".">0.72</td>
<td align="char" char=".">0.73</td>
<td align="char" char=".">0.43</td>
<td align="char" char=".">0.63</td>
</tr>
<tr>
	<td/>
<td>Over</td>
<td align="char" char=".">0.12</td>
<td align="char" char=".">0.11</td>
<td align="char" char=".">0.00</td>
<td align="char" char=".">0.04</td>
<td align="char" char=".">0.30</td>
<td align="char" char=".">0.10</td>
<td align="char" char=".">0.13</td>
<td align="char" char=".">0.10</td>
<td align="char" char=".">0.13</td>
<td align="char" char=".">0.11</td>
<td align="char" char=".">0.12</td>
<td align="char" char=".">0.01</td>
<td align="char" char=".">0.00</td>
<td align="char" char=".">0.33</td>
<td align="char" char=".">0.11</td>
</tr>
	<tr style="grey-border-top">
<td><inline-formula id="ieqn-112"><mml:math id="mml-ieqn-112"><mml:msub><mml:mrow><mml:mi mathvariant="normal">B</mml:mi><mml:mi mathvariant="normal">I</mml:mi><mml:mi mathvariant="normal">C</mml:mi></mml:mrow><mml:mrow><mml:mi>N</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula></td>
<td>Under</td>
<td align="char" char=".">0.17</td>
<td align="char" char=".">0.70</td>
<td align="char" char=".">0.74</td>
<td align="char" char=".">0.42</td>
<td align="char" char=".">0.15</td>
<td align="char" char=".">0.49</td>
<td align="char" char=".">0.38</td>
<td align="char" char=".">0.57</td>
<td align="char" char=".">0.30</td>
<td align="char" char=".">0.36</td>
<td align="char" char=".">0.51</td>
<td align="char" char=".">0.44</td>
<td align="char" char=".">0.44</td>
<td align="char" char=".">0.43</td>
<td align="char" char=".">0.44</td>
</tr>
<tr>
	<td/>		
<td>Correct</td>
<td align="char" char=".">0.81</td>
<td align="char" char=".">0.28</td>
<td align="char" char=".">0.26</td>
<td align="char" char=".">0.58</td>
<td align="char" char=".">0.79<sup>a</sup></td>
<td align="char" char=".">0.50</td>
<td align="char" char=".">0.59</td>
<td align="char" char=".">0.42</td>
<td align="char" char=".">0.66</td>
<td align="char" char=".">0.62</td>
<td align="char" char=".">0.47</td>
<td align="char" char=".">0.56</td>
<td align="char" char=".">0.56</td>
<td align="char" char=".">0.51</td>
<td align="char" char=".">0.54</td>
</tr>
<tr>
	<td/>
<td>Over</td>
<td align="char" char=".">0.03</td>
<td align="char" char=".">0.02</td>
<td align="char" char=".">0.00</td>
<td align="char" char=".">0.00</td>
<td align="char" char=".">0.06</td>
<td align="char" char=".">0.01</td>
<td align="char" char=".">0.03</td>
<td align="char" char=".">0.01</td>
<td align="char" char=".">0.03</td>
<td align="char" char=".">0.02</td>
<td align="char" char=".">0.02</td>
<td align="char" char=".">0.00</td>
<td align="char" char=".">0.00</td>
<td align="char" char=".">0.06</td>
<td align="char" char=".">0.02</td>
</tr>
<tr style="grey-border-top">
<td>Chull</td>
<td>Under</td>
<td align="char" char=".">0.00</td>
<td align="char" char=".">0.20</td>
<td align="char" char=".">0.14</td>
<td align="char" char=".">0.08</td>
<td align="char" char=".">0.08</td>
<td align="char" char=".">0.10</td>
<td align="char" char=".">0.10</td>
<td align="char" char=".">0.12</td>
<td align="char" char=".">0.07</td>
<td align="char" char=".">0.06</td>
<td align="char" char=".">0.14</td>
<td align="char" char=".">0.08</td>
<td align="char" char=".">0.07</td>
<td align="char" char=".">0.15</td>
<td align="char" char=".">0.10</td>
</tr>
<tr>
	<td/>
<td>Correct</td>
	<td align="char" char=".">0.89<sup>a</sup></td>
	<td align="char" char=".">0.66<sup>a</sup></td>
	<td align="char" char=".">0.73<sup>a</sup></td>
	<td align="char" char=".">0.81<sup>a</sup></td>
<td align="char" char=".">0.78</td>
	<td align="char" char=".">0.76<sup>a</sup></td>
	<td align="char" char=".">0.78<sup>a</sup></td>
	<td align="char" char=".">0.74<sup>a</sup></td>
	<td align="char" char=".">0.80<sup>a</sup></td>
	<td align="char" char=".">0.86<sup>a</sup></td>
	<td align="char" char=".">0.68<sup>a</sup></td>
<td align="char" char=".">0.79</td>
	<td align="char" char=".">0.86<sup>a</sup></td>
	<td align="char" char=".">0.67<sup>a</sup></td>
	<td align="char" char=".">0.77<sup>a</sup></td>
</tr>
<tr>
	<td/>
<td>Over</td>
<td align="char" char=".">0.11</td>
<td align="char" char=".">0.15</td>
<td align="char" char=".">0.14</td>
<td align="char" char=".">0.11</td>
<td align="char" char=".">0.13</td>
<td align="char" char=".">0.14</td>
<td align="char" char=".">0.12</td>
<td align="char" char=".">0.13</td>
<td align="char" char=".">0.12</td>
<td align="char" char=".">0.08</td>
<td align="char" char=".">0.18</td>
<td align="char" char=".">0.13</td>
<td align="char" char=".">0.07</td>
<td align="char" char=".">0.18</td>
<td align="char" char=".">0.13</td>
</tr>
	<tr style="grey-border-top">
<td>ICL</td>
<td>Under</td>
<td align="char" char=".">0.10</td>
<td align="char" char=".">0.47</td>
<td align="char" char=".">0.63</td>
<td align="char" char=".">0.21</td>
<td align="char" char=".">0.02</td>
<td align="char" char=".">0.33</td>
<td align="char" char=".">0.24</td>
<td align="char" char=".">0.39</td>
<td align="char" char=".">0.18</td>
<td align="char" char=".">0.22</td>
<td align="char" char=".">0.35</td>
<td align="char" char=".">0.30</td>
<td align="char" char=".">0.29</td>
<td align="char" char=".">0.27</td>
<td align="char" char=".">0.28</td>
</tr>
<tr>
	<td/>
<td>Correct</td>
<td align="char" char=".">0.79</td>
<td align="char" char=".">0.43</td>
<td align="char" char=".">0.37</td>
<td align="char" char=".">0.77</td>
<td align="char" char=".">0.70</td>
<td align="char" char=".">0.58</td>
<td align="char" char=".">0.65</td>
<td align="char" char=".">0.53</td>
<td align="char" char=".">0.70</td>
<td align="char" char=".">0.68</td>
<td align="char" char=".">0.55</td>
<td align="char" char=".">0.70</td>
<td align="char" char=".">0.71</td>
<td align="char" char=".">0.43</td>
<td align="char" char=".">0.61</td>
</tr>
<tr>
	<td/>
<td>Over</td>
<td align="char" char=".">0.11</td>
<td align="char" char=".">0.10</td>
<td align="char" char=".">0.00</td>
<td align="char" char=".">0.03</td>
<td align="char" char=".">0.28</td>
<td align="char" char=".">0.09</td>
<td align="char" char=".">0.11</td>
<td align="char" char=".">0.08</td>
<td align="char" char=".">0.12</td>
<td align="char" char=".">0.10</td>
<td align="char" char=".">0.10</td>
<td align="char" char=".">0.00</td>
<td align="char" char=".">0.00</td>
<td align="char" char=".">0.30</td>
<td align="char" char=".">0.10</td>
</tr>
	<tr style="grey-border-top">
<td>Total</td>
<td>Under</td>
<td align="char" char=".">0.07</td>
<td align="char" char=".">0.43</td>
<td align="char" char=".">0.51</td>
<td align="char" char=".">0.20</td>
<td align="char" char=".">0.05</td>
<td align="char" char=".">0.29</td>
<td align="char" char=".">0.21</td>
<td align="char" char=".">0.33</td>
<td align="char" char=".">0.16</td>
<td align="char" char=".">0.19</td>
<td align="char" char=".">0.31</td>
<td align="char" char=".">0.25</td>
<td align="char" char=".">0.25</td>
<td align="char" char=".">0.24</td>
<td align="char" char=".">0.25</td>
</tr>
<tr>
	<td/>
<td>Correct</td>
<td align="char" char=".">0.82</td>
<td align="char" char=".">0.46</td>
<td align="char" char=".">0.47</td>
<td align="char" char=".">0.74</td>
<td align="char" char=".">0.72</td>
<td align="char" char=".">0.61</td>
<td align="char" char=".">0.67</td>
<td align="char" char=".">0.57</td>
<td align="char" char=".">0.72</td>
<td align="char" char=".">0.72</td>
<td align="char" char=".">0.57</td>
<td align="char" char=".">0.72</td>
<td align="char" char=".">0.74</td>
<td align="char" char=".">0.47</td>
<td align="char" char=".">0.64</td>
</tr>
<tr>
	<td/>	
<td>Over</td>
<td align="char" char=".">0.11</td>
<td align="char" char=".">0.11</td>
<td align="char" char=".">0.02</td>
<td align="char" char=".">0.06</td>
<td align="char" char=".">0.24</td>
<td align="char" char=".">0.10</td>
<td align="char" char=".">0.12</td>
<td align="char" char=".">0.10</td>
<td align="char" char=".">0.12</td>
<td align="char" char=".">0.10</td>
<td align="char" char=".">0.12</td>
<td align="char" char=".">0.03</td>
<td align="char" char=".">0.01</td>
<td align="char" char=".">0.28</td>
<td align="char" char=".">0.11</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
	<p><italic>Note</italic>. <italic>K</italic> is the number of clusters, <italic>N</italic><sub><italic>g</italic></sub> is the within-group sample size, <italic>G</italic> is the number of groups, β is the size of the regression coefficients, <italic>Bal</italic> is balanced, <italic>Unb</italic> is unbalanced, and <inline-formula id="ieqn-113"><mml:math id="mml-ieqn-113"><mml:msub><mml:mi>σ</mml:mi><mml:mrow><mml:mi>β</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>.</p><p><sup>a</sup>Denotes best results.</p>
</table-wrap-foot>
</table-wrap>
<p>The within-group sample size <italic>N</italic><sub><italic>g</italic></sub> was most influential on the model selection performance. Specifically, on average, the model selection measures were correct 47% and 74% of the times when <inline-formula id="ieqn-114"><mml:math id="mml-ieqn-114"><mml:msub><mml:mi>N</mml:mi><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mn>50</mml:mn></mml:math></inline-formula> and <inline-formula id="ieqn-115"><mml:math id="mml-ieqn-115"><mml:msub><mml:mi>N</mml:mi><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mn>100</mml:mn></mml:math></inline-formula>, respectively. Such dramatic improvement did not hold when <italic>N</italic><sub><italic>g</italic></sub> increased to 200, for which the measures were correct 72% of the time. The improvement from <inline-formula id="ieqn-116"><mml:math id="mml-ieqn-116"><mml:msub><mml:mi>N</mml:mi><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mn>50</mml:mn></mml:math></inline-formula> to <inline-formula id="ieqn-117"><mml:math id="mml-ieqn-117"><mml:msub><mml:mi>N</mml:mi><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mn>100</mml:mn></mml:math></inline-formula> aligns with common sample size requirements in SEM, since 100 is considered the minimum for consistent estimates (<xref ref-type="bibr" rid="r17">Gorsuch, 1983</xref>). The slight decrease in performance when <inline-formula id="ieqn-118"><mml:math id="mml-ieqn-118"><mml:msub><mml:mi>N</mml:mi><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mn>200</mml:mn></mml:math></inline-formula> can be explained by the trends of under- and over-selection of the number of clusters. Generally, when choosing the incorrect model, the measures tended to under-select. However, over-selection was more prominent in the case of a large within-group sample size <inline-formula id="ieqn-119"><mml:math id="mml-ieqn-119"><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi>N</mml:mi><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mn>200</mml:mn><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula>. Such results are unsurprising, considering that larger sample sizes give more power to identify smaller differences (leading to more clusters). This is more likely in case of larger within-cluster differences <inline-formula id="ieqn-120"><mml:math id="mml-ieqn-120"><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi>σ</mml:mi><mml:mrow><mml:mi>β</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mn>0.1</mml:mn><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula>, which can be identified as additional clusters. This can be clearly seen in <xref ref-type="fig" rid="fig4">Figure 4</xref>, where the model selection performance dramatically drops when <inline-formula id="ieqn-121"><mml:math id="mml-ieqn-121"><mml:msub><mml:mi>N</mml:mi><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mn>200</mml:mn></mml:math></inline-formula> and <inline-formula id="ieqn-122"><mml:math id="mml-ieqn-122"><mml:msub><mml:mi>σ</mml:mi><mml:mrow><mml:mi>β</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mn>0.1</mml:mn></mml:math></inline-formula>.</p><fig id="fig4" position="anchor" orientation="portrait"><label>Figure  4</label><caption><title>Proportion of Correctly Selected Models in Function of the Within-Group Sample Size <italic>N</italic><sub><italic>g</italic></sub> and the Within-Cluster Differences <inline-formula id="ieqn-123"><mml:math id="mml-ieqn-123"><mml:msub><mml:mi>σ</mml:mi><mml:mrow><mml:mi>β</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula></title></caption><graphic mimetype="image" mime-subtype="png" xlink:href="meth.14931-f4.png" position="anchor" orientation="portrait"/></fig>
<p>The model selection performance was also greatly affected by the number of clusters <italic>K</italic>. On average, all model selection measures found it more difficult to identify the correct model when more clusters were underlying the data. From <xref ref-type="table" rid="t2">Table 2</xref>, the proportion of correctly selected models was better when <italic>K</italic> = 2 (0.82) than when <italic>K</italic> = 4 (0.46). Such results are in line with previous research where an increase in the true number of clusters dramatically decreased the performance of the model selection measures (<xref ref-type="bibr" rid="r8">Bulteel et al., 2013</xref>; <xref ref-type="bibr" rid="r24">Lukočienė et al., 2010</xref>) even when the cluster separation is high (<xref ref-type="bibr" rid="r36">Steinley &amp; Brusco, 2011</xref>). Similarly, the effect of the cluster size on model selection showed a comparable trend. On average, the model selection measures were correct 72% of the time when the cluster size was balanced and 57% when it was unbalanced. The effect of both the number of clusters and cluster size can be explained by lower within-cluster sample sizes in case of more and/or unbalanced clusters, which reduced the power to detect the appropriate model. The model selection was also, to a lesser extent, affected by the regression coefficients β. Specifically, a lower regression coefficient and unbalanced cluster sizes lowered the proportion of correctly selected models.</p>
<p>The interaction between two of the most important factors can be seen in <xref ref-type="fig" rid="fig5">Figure 5</xref>. The plot clearly shows that more clusters and large within-cluster differences led to a dramatic decrease in the performance of all model selection measures. Specifically, the AIC, <inline-formula id="ieqn-124"><mml:math id="mml-ieqn-124"><mml:msub><mml:mrow><mml:mi mathvariant="normal">A</mml:mi><mml:mi mathvariant="normal">I</mml:mi><mml:mi mathvariant="normal">C</mml:mi></mml:mrow><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msub></mml:math></inline-formula>, <inline-formula id="ieqn-125"><mml:math id="mml-ieqn-125"><mml:msub><mml:mrow><mml:mi mathvariant="normal">B</mml:mi><mml:mi mathvariant="normal">I</mml:mi><mml:mi mathvariant="normal">C</mml:mi></mml:mrow><mml:mrow><mml:mi>G</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, and ICL presented a substantial decrease of the correctly selected number of clusters. In contrast, the CHull and <inline-formula id="ieqn-126"><mml:math id="mml-ieqn-126"><mml:msub><mml:mrow><mml:mi mathvariant="normal">B</mml:mi><mml:mi mathvariant="normal">I</mml:mi><mml:mi mathvariant="normal">C</mml:mi></mml:mrow><mml:mrow><mml:mi>N</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> presented a lower decrease in performance when <italic>K</italic> = 4 and/or <inline-formula id="ieqn-127"><mml:math id="mml-ieqn-127"><mml:msub><mml:mi>σ</mml:mi><mml:mrow><mml:mi>β</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mn>0.1</mml:mn></mml:math></inline-formula>. <inline-formula id="ieqn-128"><mml:math id="mml-ieqn-128"><mml:msub><mml:mrow><mml:mi mathvariant="normal">B</mml:mi><mml:mi mathvariant="normal">I</mml:mi><mml:mi mathvariant="normal">C</mml:mi></mml:mrow><mml:mrow><mml:mi>N</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>'s performance was generally worse than all other model selection measures, however.</p><fig id="fig5" position="anchor" orientation="portrait"><label>Figure 5</label><caption><title>Proportion of Correctly Selected Models in Function of the Number of Clusters <italic>K</italic> and the Within-Cluster Differences <inline-formula id="ieqn-129"><mml:math id="mml-ieqn-129"><mml:msub><mml:mi>σ</mml:mi><mml:mrow><mml:mi>β</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula></title></caption><graphic mimetype="image" mime-subtype="png" xlink:href="meth.14931-f5.png" position="anchor" orientation="portrait"/></fig>
<p>From <xref ref-type="fig" rid="fig4">Figures 4</xref> and <xref ref-type="fig" rid="fig5">5</xref>, we also learn that the performance sometimes improved when going from <inline-formula id="ieqn-130"><mml:math id="mml-ieqn-130"><mml:msub><mml:mi>σ</mml:mi><mml:mrow><mml:mi>β</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mn>0</mml:mn></mml:math></inline-formula> to <inline-formula id="ieqn-131"><mml:math id="mml-ieqn-131"><mml:msub><mml:mi>σ</mml:mi><mml:mrow><mml:mi>β</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mn>0.05</mml:mn></mml:math></inline-formula>, especially for the CHull. For the CHull, this can be explained by the saturation effect, which happens when adding more clusters results in a negligible increase in the log-likelihood. This can lead to an artificially large scree ratio (because the denominator approaches zero, see <xref ref-type="disp-formula" rid="eqn-13">Equation 13</xref>), whereas, when looking at the scree plot, a virtually horizontal line, rather than an elbow, is visible at this point. In empirical practice, one can remedy this problem by looking for a clear elbow in the scree plot instead of just relying on the scree ratios.</p>
<p>The comparison between CHull and the other measures can be considered unfair, since CHull selects a model with at least two clusters. Thus, when <italic>K</italic> = 2, the other model selection measures may select a one-cluster model when the cluster separation is low, while CHull will select at least two clusters <xref ref-type="fn" rid="fn7"><sup>7</sup></xref><fn id="fn7"><label>7</label>
<p>As can be seen in <xref ref-type="table" rid="t2">Table 2</xref>, CHull never under-selects the number of clusters when <italic>K</italic> = 2</p></fn>. For a fairer comparison, we also checked the results for the other model selection measures when considering only the models from two to six clusters. In this case, AIC, <inline-formula id="ieqn-132"><mml:math id="mml-ieqn-132"><mml:msub><mml:mrow><mml:mi mathvariant="normal">A</mml:mi><mml:mi mathvariant="normal">I</mml:mi><mml:mi mathvariant="normal">C</mml:mi></mml:mrow><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msub></mml:math></inline-formula>, <inline-formula id="ieqn-133"><mml:math id="mml-ieqn-133"><mml:msub><mml:mrow><mml:mi mathvariant="normal">B</mml:mi><mml:mi mathvariant="normal">I</mml:mi><mml:mi mathvariant="normal">C</mml:mi></mml:mrow><mml:mrow><mml:mi>G</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, ICL, and <inline-formula id="ieqn-134"><mml:math id="mml-ieqn-134"><mml:msub><mml:mrow><mml:mi mathvariant="normal">B</mml:mi><mml:mi mathvariant="normal">I</mml:mi><mml:mi mathvariant="normal">C</mml:mi></mml:mrow><mml:mrow><mml:mi>N</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> selected the right number of clusters for 66.9%, 66.9%, 66%, 66.4%, and 62.8% of the datasets, respectively. Thus, their performance was closer to (but still lower than) that of the CHull (77%).</p>
<p>Finally, for a more comprehensive understanding of the outcomes, we examined how often the model selection measures correctly identified the number of clusters when considering the two best models (e.g., how often is the correct model among the two models with the lowest AIC values). The correct model was among the two best models 78.5%, 72.7%, 73.1%, 72.9%, 73.1%, and 69.3% of the times for CHull, AIC, <inline-formula id="ieqn-135"><mml:math id="mml-ieqn-135"><mml:msub><mml:mrow><mml:mi mathvariant="normal">A</mml:mi><mml:mi mathvariant="normal">I</mml:mi><mml:mi mathvariant="normal">C</mml:mi></mml:mrow><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msub></mml:math></inline-formula>, <inline-formula id="ieqn-136"><mml:math id="mml-ieqn-136"><mml:msub><mml:mrow><mml:mi mathvariant="normal">B</mml:mi><mml:mi mathvariant="normal">I</mml:mi><mml:mi mathvariant="normal">C</mml:mi></mml:mrow><mml:mrow><mml:mi>G</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, ICL, and <inline-formula id="ieqn-137"><mml:math id="mml-ieqn-137"><mml:msub><mml:mrow><mml:mi mathvariant="normal">B</mml:mi><mml:mi mathvariant="normal">I</mml:mi><mml:mi mathvariant="normal">C</mml:mi></mml:mrow><mml:mrow><mml:mi>N</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, respectively. The <inline-formula id="ieqn-138"><mml:math id="mml-ieqn-138"><mml:msub><mml:mrow><mml:mi mathvariant="normal">B</mml:mi><mml:mi mathvariant="normal">I</mml:mi><mml:mi mathvariant="normal">C</mml:mi></mml:mrow><mml:mrow><mml:mi>G</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, <inline-formula id="ieqn-139"><mml:math id="mml-ieqn-139"><mml:msub><mml:mrow><mml:mi mathvariant="normal">B</mml:mi><mml:mi mathvariant="normal">I</mml:mi><mml:mi mathvariant="normal">C</mml:mi></mml:mrow><mml:mrow><mml:mi>N</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, and ICL showed the largest improvements in performance compared to when we focus only on the best model.</p></sec>
<sec><title>Cluster Recovery</title>
<p>The model selection results must be interpreted in light of the cluster recovery (i.e., to what extent MMG-SEM assigns groups to the correct clusters when the correct number of clusters is specified). To this end, we computed the Adjusted Rand Index (ARI; <xref ref-type="bibr" rid="r21">Hubert &amp; Arabie, 1985</xref>) for the model with the true number of clusters. For instance, if <italic>K</italic> = 2, we computed the ARI for the model with two clusters (regardless of what the model selection measures chose). The ARI compares two partitions (i.e., modal assignments of the groups to a cluster), taking on a value of 1 for complete agreement and 0 when agreement does not exceed that between two random partitions. Note that it can take on negative values if the agreement is less than what is expected at random. The average ARI across all conditions was good (0.93), but it ranged from -0.10 to 1. Per model selection measure, we also inspected the average ARI across data sets depending on whether the number of clusters was under-, over-, or correctly selected (<xref ref-type="table" rid="t3">Table 3</xref>). Clearly, selecting the incorrect model (i.e., under- or over-selection) was related to the ARI being lower for the correct model. Specifically, the average ARI was below 0.89 for all measures when the selected model was incorrect, while it was above 0.97 when the correct model was selected.</p>
<table-wrap id="t3" position="anchor" orientation="portrait">
<label>Table 3</label><caption><title>Average ARI for the Model With the True Number of Clusters Depending On Whether the Number of Clusters Was Correctly Selected or Over- or Under-Selected by the Model Selection Measures</title></caption>
<table frame="hsides" rules="groups" style="compact-1"><colgroup span="1">
	<col width="" align="left"/>
	<col width="" align="left"/>
	<col width="" align="left"/>
	<col width="" align="left"/>
	<col width="" align="left"/>
	<col width="" align="left"/>
	<col width="" align="left"/></colgroup>
<thead>
<tr>
<th>Result</th>
<th>AIC</th>
<th><inline-formula id="ieqn-140"><mml:math id="mml-ieqn-140"><mml:msub><mml:mrow><mml:mi mathvariant="normal">A</mml:mi><mml:mi mathvariant="normal">I</mml:mi><mml:mi mathvariant="normal">C</mml:mi></mml:mrow><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msub></mml:math></inline-formula></th>
<th><inline-formula id="ieqn-141"><mml:math id="mml-ieqn-141"><mml:msub><mml:mrow><mml:mi mathvariant="normal">B</mml:mi><mml:mi mathvariant="normal">I</mml:mi><mml:mi mathvariant="normal">C</mml:mi></mml:mrow><mml:mrow><mml:mi>G</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula></th>
<th><inline-formula id="ieqn-142"><mml:math id="mml-ieqn-142"><mml:msub><mml:mrow><mml:mi mathvariant="normal">B</mml:mi><mml:mi mathvariant="normal">I</mml:mi><mml:mi mathvariant="normal">C</mml:mi></mml:mrow><mml:mrow><mml:mi>N</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula></th>
<th>CHull</th>
<th>ICL</th>
</tr>
</thead>
<tbody>
<tr>
<td>Under</td>
<td>0.83 (0.19)</td>
<td>0.85 (0.19)</td>
<td>0.86 (0.19)</td>
<td>0.89 (0.18)</td>
<td>0.77 (0.22)</td>
<td>0.87 (0.18)</td>
</tr>
<tr>
<td>Correct</td>
<td>0.98 (0.07)</td>
<td>0.98 (0.07)</td>
<td>0.98 (0.07)</td>
<td>0.98 (0.08)</td>
<td>0.97 (0.09)</td>
<td>0.98 (0.07)</td>
</tr>
<tr>
<td>Over</td>
<td>0.86 (0.20)</td>
<td>0.85 (0.20)</td>
<td>0.85 (0.20)</td>
<td>0.89 (0.18)</td>
<td>0.85 (0.20)</td>
<td>0.85 (0.21)</td>
</tr>
</tbody>
</table>
	<table-wrap-foot>
		<p><italic>Note</italic>. The standard deviation is in parentheses.</p>
	</table-wrap-foot>	
</table-wrap></sec>
<sec><title>Section Conclusion</title>
	<p>Before drawing conclusions, it is important to note that, strictly speaking, there are as many clusters as there are groups when <inline-formula id="ieqn-144"><mml:math id="mml-ieqn-144"><mml:msub><mml:mi>σ</mml:mi><mml:mrow><mml:mi>β</mml:mi></mml:mrow></mml:msub><mml:mo>&gt;</mml:mo><mml:mn>0</mml:mn></mml:math></inline-formula> (i.e., in case of within-cluster differences). As MMG-SEM does not capture within-cluster differences, the model selection could suggest extracting more clusters or even capturing each group as a separate cluster, especially when there is enough power to find small differences. However, in this study, we still assumed the true number of clusters to be <italic>K</italic>, since it is desirable to assign groups with very similar regression parameters to the same cluster (see <xref ref-type="sec" rid="intro	">Introduction</xref>). Considering every group as a separate cluster would boil back down to an MG-SEM with group-specific relations and the pesky pairwise comparisons thereof. As this is what we wanted to avoid, we did not include the MG-SEM model in this study.</p>
<p>In an extensive simulation study, we assessed six different model selection measures for MMG-SEM. As expected, lower within-group sample sizes <inline-formula id="ieqn-145"><mml:math id="mml-ieqn-145"><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi>N</mml:mi><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mn>50</mml:mn><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula>, large within-cluster differences <inline-formula id="ieqn-146"><mml:math id="mml-ieqn-146"><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi>σ</mml:mi><mml:mrow><mml:mi>β</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mn>0.1</mml:mn><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula>, more clusters (<italic>K</italic> = 4), and unbalanced cluster sizes decreased the performance of all measures. Overall, the best-performing measure was the CHull and the worst was the <inline-formula id="ieqn-147"><mml:math id="mml-ieqn-147"><mml:msub><mml:mrow><mml:mi mathvariant="normal">B</mml:mi><mml:mi mathvariant="normal">I</mml:mi><mml:mi mathvariant="normal">C</mml:mi></mml:mrow><mml:mrow><mml:mi>N</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>. AIC, <inline-formula id="ieqn-148"><mml:math id="mml-ieqn-148"><mml:mrow><mml:mi mathvariant="normal">A</mml:mi><mml:mi mathvariant="normal">I</mml:mi><mml:msub><mml:mi mathvariant="normal">C</mml:mi><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:math></inline-formula>, <inline-formula id="ieqn-149"><mml:math id="mml-ieqn-149"><mml:msub><mml:mrow><mml:mi mathvariant="normal">B</mml:mi><mml:mi mathvariant="normal">I</mml:mi><mml:mi mathvariant="normal">C</mml:mi></mml:mrow><mml:mrow><mml:mi>G</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, and ICL presented similar performances with minor differences depending on specific conditions.</p>
<p>Considering our results, we suggest using the CHull when performing model selection for MMG-SEM. Since it cannot select the minimum or maximum number of clusters, we suggest also inspecting CHull’s scree plot and looking for an elbow to confirm the number of clusters. If no elbow is visible, the most appropriate number of clusters is likely one or the maximum number evaluated. Furthermore, since CHull was not universally best across all data sets<xref ref-type="fn" rid="fn8"><sup>8</sup></xref><fn id="fn8"><label>8</label>
<p>For instance, out of the total 14400 data sets: (1) AIC and CHull were both correct in 8377 cases; (2) AIC was incorrect and CHull correct in 2754 cases; and (3) AIC was correct and CHull incorrect in 1105 cases.</p></fn>, we recommend combining its results with at least one of the other measures (e.g., AIC, <inline-formula id="ieqn-150"><mml:math id="mml-ieqn-150"><mml:mrow><mml:mi mathvariant="normal">A</mml:mi><mml:mi mathvariant="normal">I</mml:mi><mml:msub><mml:mi mathvariant="normal">C</mml:mi><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:math></inline-formula> or <inline-formula id="ieqn-151"><mml:math id="mml-ieqn-151"><mml:msub><mml:mrow><mml:mi mathvariant="normal">B</mml:mi><mml:mi mathvariant="normal">I</mml:mi><mml:mi mathvariant="normal">C</mml:mi></mml:mrow><mml:mrow><mml:mi>G</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>) to validate a decision. For instance, selecting <italic>K</italic> = 1 when no elbow is found for CHull and AIC suggests a one-cluster model (a small simulation study about model selection performance when <italic>K</italic> = 1 can be found in <xref ref-type="bibr" rid="r31.5">Perez Alonso et al., 2025</xref>). It may also help to consider the best solution for the different measures and compare the solutions in terms of which differences in structural relations are found (and which clustering) and how this relates to prior theories and previous research about the matter.</p></sec></sec>
<sec sec-type="General_Discussion"><title>General Discussion</title>
<p>When using MMG-SEM, the user must specify the appropriate number of clusters for a given data set, as is the case for all clustering techniques. However, the ‘true’ number of clusters is typically unknown when dealing with real-world data. Therefore, researchers often rely on model selection measures to decide on the number of clusters. Several model selection measures have been evaluated for other clustering methods, but there is no clear-cut answer to which measure is the best one. Different results were found depending on the clustering method, the conditions assessed, and the level at which the clustering is performed (i.e., observation or group level) (<xref ref-type="bibr" rid="r2">Akogul &amp; Erisoglu, 2016</xref>; <xref ref-type="bibr" rid="r16">De Roover et al., 2022</xref>; <xref ref-type="bibr" rid="r24">Lukočienė et al., 2010</xref>; <xref ref-type="bibr" rid="r30">Nylund et al., 2007</xref>). Considering the conflicting results and the unique properties of MMG-SEM, such as the combination of group- and cluster-specific parameters, and a clustering focused on regression parameters, prior conclusions on their model selection performance cannot be generalized to MMG-SEM. Therefore, this paper aimed to find the best-performing model selection measure for MMG-SEM through an extensive simulation study. In particular, we compared six model selection measures (i.e., CHull, AIC, <inline-formula id="ieqn-152"><mml:math id="mml-ieqn-152"><mml:msub><mml:mrow><mml:mi mathvariant="normal">A</mml:mi><mml:mi mathvariant="normal">I</mml:mi><mml:mi mathvariant="normal">C</mml:mi></mml:mrow><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msub></mml:math></inline-formula>, <inline-formula id="ieqn-153"><mml:math id="mml-ieqn-153"><mml:msub><mml:mrow><mml:mi mathvariant="normal">B</mml:mi><mml:mi mathvariant="normal">I</mml:mi><mml:mi mathvariant="normal">C</mml:mi></mml:mrow><mml:mrow><mml:mi>G</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, ICL, and <inline-formula id="ieqn-154"><mml:math id="mml-ieqn-154"><mml:msub><mml:mrow><mml:mi mathvariant="normal">B</mml:mi><mml:mi mathvariant="normal">I</mml:mi><mml:mi mathvariant="normal">C</mml:mi></mml:mrow><mml:mrow><mml:mi>N</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>), and included conditions that affect the cluster’s separability and mimic empirically realistic conditions. In particular, the small within-cluster differences resembled the small (and trivial) differences between groups that will often be found in empirical research but that are ineffectual to the substantive conclusions on how structural relations differ.</p>
<p>Overall, the best-performing measure was the CHull, followed by the AIC, <inline-formula id="ieqn-155"><mml:math id="mml-ieqn-155"><mml:msub><mml:mrow><mml:mi mathvariant="normal">A</mml:mi><mml:mi mathvariant="normal">I</mml:mi><mml:mi mathvariant="normal">C</mml:mi></mml:mrow><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msub></mml:math></inline-formula>, <inline-formula id="ieqn-156"><mml:math id="mml-ieqn-156"><mml:msub><mml:mrow><mml:mi mathvariant="normal">B</mml:mi><mml:mi mathvariant="normal">I</mml:mi><mml:mi mathvariant="normal">C</mml:mi></mml:mrow><mml:mrow><mml:mi>G</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, ICL, and <inline-formula id="ieqn-157"><mml:math id="mml-ieqn-157"><mml:msub><mml:mrow><mml:mi mathvariant="normal">B</mml:mi><mml:mi mathvariant="normal">I</mml:mi><mml:mi mathvariant="normal">C</mml:mi></mml:mrow><mml:mrow><mml:mi>N</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>. While, in general, this is in line with some previous studies (e.g., <xref ref-type="bibr" rid="r8">Bulteel et al., 2013</xref>; <xref ref-type="bibr" rid="r15">De Roover, 2021</xref>; <xref ref-type="bibr" rid="r16">De Roover et al., 2022</xref>), the clear difference in performance between CHull and the other measures was a remarkable find. This difference could not be explained by the fact that CHull can only select at least two clusters. CHull’s advantage may result from its flexibility and lack of assumptions compared to the other measures. For instance, some argue that the true model must be among the candidates for BIC to have consistent results (<xref ref-type="bibr" rid="r41">Vrieze, 2012</xref>), which was, strictly speaking, not always the case in the simulation study since the ‘true’ model was one with group-specific (instead of cluster-specific) regression parameters in case of within-cluster differences.</p>
<p>The vastly inferior performance of <inline-formula id="ieqn-158"><mml:math id="mml-ieqn-158"><mml:msub><mml:mrow><mml:mi mathvariant="normal">B</mml:mi><mml:mi mathvariant="normal">I</mml:mi><mml:mi mathvariant="normal">C</mml:mi></mml:mrow><mml:mrow><mml:mi>N</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> coincides with previous results for mixture models at the group level that showed that using the number of observations <italic>N</italic> instead of <italic>G</italic> as the sample size in the BIC leads to overpenalization and, thus, under-selection (<xref ref-type="bibr" rid="r15">De Roover, 2021</xref>; <xref ref-type="bibr" rid="r16">De Roover et al., 2022</xref>; <xref ref-type="bibr" rid="r24">Lukočienė et al., 2010</xref>; <xref ref-type="bibr" rid="r25">Lukočienė &amp; Vermunt, 2009</xref>). More surprising was the superior performance of the AIC over the <inline-formula id="ieqn-159"><mml:math id="mml-ieqn-159"><mml:msub><mml:mrow><mml:mi mathvariant="normal">B</mml:mi><mml:mi mathvariant="normal">I</mml:mi><mml:mi mathvariant="normal">C</mml:mi></mml:mrow><mml:mrow><mml:mi>G</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, given that previous simulations have shown a slight advantage of <inline-formula id="ieqn-160"><mml:math id="mml-ieqn-160"><mml:msub><mml:mrow><mml:mi mathvariant="normal">B</mml:mi><mml:mi mathvariant="normal">I</mml:mi><mml:mi mathvariant="normal">C</mml:mi></mml:mrow><mml:mrow><mml:mi>G</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> in latent class and mixture models (<xref ref-type="bibr" rid="r15">De Roover, 2021</xref>; <xref ref-type="bibr" rid="r24">Lukočienė et al., 2010</xref>; <xref ref-type="bibr" rid="r25">Lukočienė &amp; Vermunt, 2009</xref>). In other studies, AIC slightly outperformed the <inline-formula id="ieqn-161"><mml:math id="mml-ieqn-161"><mml:msub><mml:mrow><mml:mi mathvariant="normal">B</mml:mi><mml:mi mathvariant="normal">I</mml:mi><mml:mi mathvariant="normal">C</mml:mi></mml:mrow><mml:mrow><mml:mi>G</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, however (<xref ref-type="bibr" rid="r16">De Roover et al., 2022</xref>).</p>
	<p>It is worth mentioning that simulation-based research always comes with limitations. Specifically, the results cannot be straightforwardly generalized to conditions that were not assessed in the simulation. For instance, we only included data and models with continuous and normally distributed items, whereas earlier simulations showed different results depending on the type of indicator (<xref ref-type="bibr" rid="r24">Lukočienė et al., 2010</xref>). Since data in social sciences commonly uses ordinal indicators and often presents non-normality (e.g., skewness), it is important to extend the model selection evaluations to MMG-SEM with ordinal indicators and robust estimators for non-normality. Note that the CHull could be computed with measures of model fit other than log-likelihood (e.g., the distance between the model-implied and observed covariance matrices) and avoid the distributional assumptions that come with it. Moreover, in the <xref ref-type="sec" rid="sim_stu">Simulation Study</xref>, we focused on model selection measures that are most commonly used for mixture models in social sciences, overlooking other measures. For instance, the Kullback Information Criterion (<xref ref-type="bibr" rid="r10">Cavanaugh, 1999</xref>) is a promising alternative for large sample conditions (<xref ref-type="bibr" rid="r2">Akogul &amp; Erisoglu, 2016</xref>); the normalized information criteria (<xref ref-type="bibr" rid="r14">Cohen &amp; Berchenko, 2021</xref>) was developed for the presence of missing data; and the Likelihood Increment Percentage per Parameter (<xref ref-type="bibr" rid="r18">Grimm et al., 2021</xref>; <xref ref-type="bibr" rid="r27">McArdle et al., 2002</xref>) provides a ‘effect size’ measure of the relative fit improvement in the context of model selection, and is thus an alternative to CHull.</p>
<p>Finally, for MMG-SEM — which compares structural relations between groups using clusters — selecting an appropriate number of clusters is essential for the research questions. Indeed, under- or over-selecting clusters may impair the conclusions on differences and similarities in the relations of interest. When selecting too few clusters, important differences may be overlooked. When selecting too many clusters, one ends up with an overly complex model that implies more comparisons of cluster-specific regression coefficients and, thus, a higher risk of false positives. Therefore, we are happy to conclude that the model selection measures assessed in this paper offer promising solutions to the model selection problem for MMG-SEM, especially when combined (e.g., CHull and AIC).</p></sec>
</body>
<back>
	
	
	
<ref-list><title>References</title>
<ref id="r1"><mixed-citation publication-type="journal"><string-name name-style="western"><surname>Akaike</surname>, <given-names>H.</given-names></string-name> (<year>1974</year>). <article-title>A new look at the statistical model identification</article-title>. <source><italic>IEEE Transactions on Automatic Control</italic></source>, <volume>19</volume> (<issue>6</issue>), <fpage>716</fpage>–<lpage>723</lpage>. <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1109/TAC.1974.1100705">https://doi.org/10.1109/TAC.1974.1100705</ext-link></mixed-citation></ref>
<ref id="r2"><mixed-citation publication-type="journal"><string-name name-style="western"><surname>Akogul</surname>, <given-names>S.</given-names></string-name>, &amp; <string-name name-style="western"><surname>Erisoglu</surname>, <given-names>M.</given-names></string-name> (<year>2016</year>). <article-title>A comparison of information criteria in clustering based on mixture of multivariate normal distributions</article-title>. <source><italic>Mathematical and Computational Applications</italic></source>, <volume>21</volume> (<issue>3</issue>), <elocation-id>34</elocation-id>. <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.3390/mca21030034">https://doi.org/10.3390/mca21030034</ext-link> </mixed-citation></ref>
<ref id="r3"><mixed-citation publication-type="journal"><string-name name-style="western"><surname>Bakk</surname>, <given-names>Z.</given-names></string-name>, <string-name name-style="western"><surname>Oberski</surname>, <given-names>D. L.</given-names></string-name>, &amp; <string-name name-style="western"><surname>Vermunt</surname>, <given-names>J. K.</given-names></string-name> (<year>2014</year>). <article-title>Relating latent class assignments to external variables: Standard errors for correct inference</article-title>. <source><italic>Political Analysis</italic></source>, <volume>22</volume> (<issue>4</issue>), <fpage>520</fpage>–<lpage>540</lpage>. <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1093/pan/mpu003">https://doi.org/10.1093/pan/mpu003</ext-link></mixed-citation></ref>
<ref id="r4"><mixed-citation publication-type="journal"><string-name name-style="western"><surname>Bastian</surname>, <given-names>B.</given-names></string-name>, <string-name name-style="western"><surname>Kuppens</surname>, <given-names>P.</given-names></string-name>, <string-name name-style="western"><surname>De Roover</surname>, <given-names>K.</given-names></string-name>, &amp; <string-name name-style="western"><surname>Diener</surname>, <given-names>E.</given-names></string-name> (<year>2014</year>). <article-title>Is valuing positive emotion associated with life satisfaction?</article-title> <source><italic>Emotion</italic></source>, <volume>14</volume> (<issue>4</issue>), <fpage>639</fpage>–<lpage>645</lpage>. <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1037/a0036466">https://doi.org/10.1037/a0036466</ext-link></mixed-citation></ref>
<ref id="r5"><mixed-citation publication-type="journal"><string-name name-style="western"><surname>Biernacki</surname>, <given-names>C.</given-names></string-name>, <string-name name-style="western"><surname>Celeux</surname>, <given-names>G.</given-names></string-name>, &amp; <string-name name-style="western"><surname>Govaert</surname>, <given-names>G.</given-names></string-name> (<year>2000</year>). <article-title>Assessing a mixture model for clustering with the integrated completed likelihood</article-title>. <source><italic>IEEE Transactions on Pattern Analysis and Machine Intelligence</italic></source>, <volume>22</volume> (<issue>7</issue>), <fpage>719</fpage>–<lpage>725</lpage>. <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1109/34.865189">https://doi.org/10.1109/34.865189</ext-link></mixed-citation></ref>
<ref id="r6"><mixed-citation publication-type="book"><string-name name-style="western"><surname>Bollen</surname>, <given-names>K. A.</given-names></string-name> (<year>1989</year>). <source><italic>Structural equations with latent variables</italic></source>. <publisher-name>John Wiley &amp; Sons</publisher-name>.</mixed-citation></ref>
<ref id="r7"><mixed-citation publication-type="book"><string-name name-style="western"><surname>Bozdogan</surname>, <given-names>H.</given-names></string-name> (<year>1994</year>). Mixture-model cluster analysis using model selection criteria and a new informational measure of complexity. In H. Bozdogan, S. L. Sclove, A. K. Gupta, D. Haughton, G. Kitagawa, T. Ozaki &amp; K. Tanabe (Eds.), <italic>Proceedings of the First US/Japan Conference on the Frontiers of Statistical Modeling: An informational approach</italic> (pp. 69–113). Springer Netherlands. <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1007/978-94-011-0800-3_3">https://doi.org/10.1007/978-94-011-0800-3_3</ext-link></mixed-citation></ref>
<ref id="r8"><mixed-citation publication-type="journal"><string-name name-style="western"><surname>Bulteel</surname>, <given-names>K.</given-names></string-name>, <string-name name-style="western"><surname>Wilderjans</surname>, <given-names>T. F.</given-names></string-name>, <string-name name-style="western"><surname>Tuerlinckx</surname>, <given-names>F.</given-names></string-name>, &amp; <string-name name-style="western"><surname>Ceulemans</surname>, <given-names>E.</given-names></string-name> (<year>2013</year>). <article-title>CHull as an alternative to AIC and BIC in the context of mixtures of factor analyzers</article-title>. <source><italic>Behavior Research Methods</italic></source>, <volume>45</volume> (<issue>3</issue>), <fpage>782</fpage>–<lpage>791</lpage>. <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.3758/s13428-012-0293-y">https://doi.org/10.3758/s13428-012-0293-y</ext-link></mixed-citation></ref>
<ref id="r9"><mixed-citation publication-type="journal"><string-name name-style="western"><surname>Byrne</surname>, <given-names>B. M.</given-names></string-name>, <string-name name-style="western"><surname>Shavelson</surname>, <given-names>R. J.</given-names></string-name>, &amp; <string-name name-style="western"><surname>Muthén</surname>, <given-names>B.</given-names></string-name> (<year>1989</year>). <article-title>Testing for the equivalence of factor covariance and mean structures: The issue of partial measurement invariance</article-title>. <source><italic>Psychological Bulletin</italic></source>, <volume>105</volume> (<issue>3</issue>), <fpage>456</fpage>–<lpage>466</lpage>. <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1037/0033-2909.105.3.456">https://doi.org/10.1037/0033-2909.105.3.456</ext-link></mixed-citation></ref>
<ref id="r10"><mixed-citation publication-type="journal"><string-name name-style="western"><surname>Cavanaugh</surname>, <given-names>J. E.</given-names></string-name> (<year>1999</year>). <article-title>A large-sample model selection criterion based on Kullback’s symmetric divergence</article-title>. <source><italic>Statistics &amp; Probability Letters</italic></source>, <volume>42</volume> (<issue>4</issue>), <fpage>333</fpage>–<lpage>343</lpage>. <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1016/S0167-7152(98)00200-4">https://doi.org/10.1016/S0167-7152(98)00200-4</ext-link></mixed-citation></ref>
<ref id="r11"><mixed-citation publication-type="journal"><string-name name-style="western"><surname>Ceulemans</surname>, <given-names>E.</given-names></string-name>, &amp; <string-name name-style="western"><surname>Kiers</surname>, <given-names>H. A. L.</given-names></string-name> (<year>2006</year>). <article-title>Selecting among three-mode principal component models of different types and complexities: A numerical convex hull based method</article-title>. <source><italic>British Journal of Mathematical and Statistical Psychology</italic></source>, <volume>59</volume> (<issue>1</issue>), <fpage>133</fpage>–<lpage>150</lpage>. <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1348/000711005X64817">https://doi.org/10.1348/000711005X64817</ext-link></mixed-citation></ref>
<ref id="r12"><mixed-citation publication-type="journal"><string-name name-style="western"><surname>Chen</surname>, <given-names>F. F.</given-names></string-name> (<year>2007</year>). <article-title>Sensitivity of goodness of fit indexes to lack of measurement invariance</article-title>. <source><italic>Structural Equation Modeling: A Multidisciplinary Journal</italic></source>, <volume>14</volume> (<issue>3</issue>), <fpage>464</fpage>–<lpage>504</lpage>. <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1080/10705510701301834">https://doi.org/10.1080/10705510701301834</ext-link></mixed-citation></ref>
<ref id="r13"><mixed-citation publication-type="journal"><string-name name-style="western"><surname>Chen</surname>, <given-names>F. F.</given-names></string-name> (<year>2008</year>). <article-title>What happens if we compare chopsticks with forks? The impact of making inappropriate comparisons in cross-cultural research</article-title>. <source><italic>Journal of Personality and Social Psychology</italic></source>, <volume>95</volume> (<issue>5</issue>), <fpage>1005</fpage>–<lpage>1018</lpage>. <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1037/a0013193">https://doi.org/10.1037/a0013193</ext-link></mixed-citation></ref>
<ref id="r14"><mixed-citation publication-type="journal"><string-name name-style="western"><surname>Cohen</surname>, <given-names>N.</given-names></string-name>, &amp; <string-name name-style="western"><surname>Berchenko</surname>, <given-names>Y.</given-names></string-name> (<year>2021</year>). <article-title>Normalized information criteria and model selection in the presence of missing data</article-title>. <source><italic>Mathematics</italic></source>, <volume>9</volume> (<issue>19</issue>), <elocation-id>2474</elocation-id>. <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.3390/math9192474">https://doi.org/10.3390/math9192474</ext-link></mixed-citation></ref>
<ref id="r15"><mixed-citation publication-type="journal"><string-name name-style="western"><surname>De Roover</surname>, <given-names>K.</given-names></string-name> (<year>2021</year>). <article-title>Finding clusters of groups with measurement invariance: Unraveling intercept non-invariance with Mixture Multigroup Factor Analysis</article-title>. <source><italic>Structural Equation Modeling: A Multidisciplinary Journal</italic></source>, <volume>28</volume> (<issue>5</issue>), <fpage>663</fpage>–<lpage>683</lpage>. <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1080/10705511.2020.1866577">https://doi.org/10.1080/10705511.2020.1866577</ext-link></mixed-citation></ref>
<ref id="r16"><mixed-citation publication-type="journal"><string-name name-style="western"><surname>De Roover</surname>, <given-names>K.</given-names></string-name>, <string-name name-style="western"><surname>Vermunt</surname>, <given-names>J. K.</given-names></string-name>, &amp; <string-name name-style="western"><surname>Ceulemans</surname>, <given-names>E.</given-names></string-name> (<year>2022</year>). <article-title>Mixture multigroup factor analysis for unraveling factor loading noninvariance across many groups</article-title>. <source><italic>Psychological Methods</italic></source>, <volume>27</volume> (<issue>3</issue>), <fpage>281</fpage>–<lpage>306</lpage>. <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1037/met0000355">https://doi.org/10.1037/met0000355</ext-link></mixed-citation></ref>
<ref id="r17"><mixed-citation publication-type="book"><string-name name-style="western"><surname>Gorsuch</surname>, <given-names>R.</given-names></string-name> (<year>1983</year>). <source><italic>Factor Analysis</italic></source> (2<sup>nd</sup> ed.). <publisher-loc>Lawrence Erlbaum</publisher-loc>.</mixed-citation></ref>
<ref id="r18"><mixed-citation publication-type="journal"><string-name name-style="western"><surname>Grimm</surname>, <given-names>K. J.</given-names></string-name>, <string-name name-style="western"><surname>Houpt</surname>, <given-names>R.</given-names></string-name>, &amp; <string-name name-style="western"><surname>Rodgers</surname>, <given-names>D.</given-names></string-name> (<year>2021</year>). <article-title>Model fit and comparison in Finite Mixture Models: A review and a novel approach</article-title>. <source><italic>Frontiers in Education</italic></source>, <volume>6</volume>, <elocation-id>613645</elocation-id>. <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.3389/feduc.2021.613645">https://doi.org/10.3389/feduc.2021.613645</ext-link></mixed-citation></ref>
	<ref id="r19"><mixed-citation publication-type="journal"><string-name name-style="western"><surname>Guenole</surname>, <given-names>N.</given-names></string-name>, &amp; <string-name name-style="western"><surname>Brown</surname>, <given-names>A.</given-names></string-name> (<year>2014</year>). <article-title>The consequences of ignoring measurement invariance for path coefficients in structural equation models</article-title>. <source><italic>Frontiers in Psychology</italic></source>, <volume>5</volume>, <elocation-id>980</elocation-id>. <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.3389/fpsyg.2014.00980">https://doi.org/10.3389/fpsyg.2014.00980</ext-link></mixed-citation></ref>
<ref id="r20"><mixed-citation publication-type="book"><string-name name-style="western"><surname>Hox</surname>, <given-names>J. J.</given-names></string-name>, <string-name name-style="western"><surname>Moerbeek</surname>, <given-names>M.</given-names></string-name>, &amp; <string-name name-style="western"><surname>Van De Schoot</surname>, <given-names>R.</given-names></string-name> (<year>2017</year>). <source><italic>Multilevel analysis: Techniques and applications</italic></source> <edition>(3<sup>rd</sup> ed.)</edition>. <publisher-name>Routledge</publisher-name>.</mixed-citation></ref>
<ref id="r21"><mixed-citation publication-type="journal"><string-name name-style="western"><surname>Hubert</surname>, <given-names>L.</given-names></string-name>, &amp; <string-name name-style="western"><surname>Arabie</surname>, <given-names>P.</given-names></string-name> (<year>1985</year>). <article-title>Comparing partitions</article-title>. <source><italic>Journal of Classification</italic></source>, <volume>2</volume> (<issue>1</issue>), <fpage>193</fpage>–<lpage>218</lpage>. <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1007/BF01908075">https://doi.org/10.1007/BF01908075</ext-link></mixed-citation></ref>
<ref id="r22"><mixed-citation publication-type="journal"><string-name name-style="western"><surname>Kim</surname>, <given-names>E. S.</given-names></string-name>, <string-name name-style="western"><surname>Joo</surname>, <given-names>S.-H.</given-names></string-name>, <string-name name-style="western"><surname>Lee</surname>, <given-names>P.</given-names></string-name>, <string-name name-style="western"><surname>Wang</surname>, <given-names>Y.</given-names></string-name>, &amp; <string-name name-style="western"><surname>Stark</surname>, <given-names>S.</given-names></string-name> (<year>2016</year>). <article-title>Measurement invariance testing across between-level latent classes using Multilevel Factor Mixture Modeling</article-title>. <source><italic>Structural Equation Modeling: A Multidisciplinary Journal</italic></source>, <volume>23</volume> (<issue>6</issue>), <fpage>870</fpage>–<lpage>887</lpage>. <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1080/10705511.2016.1196108">https://doi.org/10.1080/10705511.2016.1196108</ext-link></mixed-citation></ref>
<ref id="r23"><mixed-citation publication-type="journal"><string-name name-style="western"><surname>Kuppens</surname>, <given-names>P.</given-names></string-name>, <string-name name-style="western"><surname>Realo</surname>, <given-names>A.</given-names></string-name>, &amp; <string-name name-style="western"><surname>Diener</surname>, <given-names>E.</given-names></string-name> (<year>2008</year>). <article-title>The role of positive and negative emotions in life satisfaction judgment across nations</article-title>. <source><italic>Journal of Personality and Social Psychology</italic></source>, <volume>95</volume> (<issue>1</issue>), <fpage>66</fpage>–<lpage>75</lpage>. <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1037/0022-3514.95.1.66">https://doi.org/10.1037/0022-3514.95.1.66</ext-link></mixed-citation></ref>
<ref id="r24"><mixed-citation publication-type="journal"><string-name name-style="western"><surname>Lukočienė</surname>, <given-names>O.</given-names></string-name>, <string-name name-style="western"><surname>Varriale</surname>, <given-names>R.</given-names></string-name>, &amp; <string-name name-style="western"><surname>Vermunt</surname>, <given-names>J. K.</given-names></string-name> (<year>2010</year>). <article-title>The simultaneous decision(s) about the number of lower- and higher-level classes in Multilevel Latent Class Analysis</article-title>. <source><italic>Sociological Methodology</italic></source>, <volume>40</volume> (<issue>1</issue>), <fpage>247</fpage>–<lpage>283</lpage>. <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1111/j.1467-9531.2010.01231.x">https://doi.org/10.1111/j.1467-9531.2010.01231.x</ext-link></mixed-citation></ref>
<ref id="r25"><mixed-citation publication-type="book"><string-name name-style="western"><surname>Lukočienė</surname>, <given-names>O.</given-names></string-name>, &amp; <string-name name-style="western"><surname>Vermunt</surname>, <given-names>J. K.</given-names></string-name> (<year>2009</year>). <chapter-title>Determining the number of components in mixture models for hierarchical data</chapter-title>. In A. Fink, B. Lausen, W. Seidel &amp; A. Ultsch  (Eds.), <italic>Advances in data analysis, data handling and business intelligence</italic> (pp. 241–249). Springer Berlin Heidelberg. <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1007/978-3-642-01044-6_22">https://doi.org/10.1007/978-3-642-01044-6_22</ext-link></mixed-citation></ref>
<ref id="r26"><mixed-citation publication-type="journal"><string-name name-style="western"><surname>Mayerl</surname>, <given-names>J.</given-names></string-name>, &amp; <string-name name-style="western"><surname>Best</surname>, <given-names>H.</given-names></string-name> (<year>2019</year>). <article-title>Attitudes and behavioral intentions to protect the environment: How consistent is the structure of environmental concern in cross-national comparison?</article-title> <source><italic>International Journal of Sociology</italic></source>, <volume>49</volume> (<issue>1</issue>), <fpage>27</fpage>–<lpage>52</lpage>. <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1080/00207659.2018.1560980">https://doi.org/10.1080/00207659.2018.1560980</ext-link></mixed-citation></ref>
<ref id="r27"><mixed-citation publication-type="journal"><string-name name-style="western"><surname>McArdle</surname>, <given-names>J. J.</given-names></string-name>, <string-name name-style="western"><surname>Ferrer-Caja</surname>, <given-names>E.</given-names></string-name>, <string-name name-style="western"><surname>Hamagami</surname>, <given-names>F.</given-names></string-name>, &amp; <string-name name-style="western"><surname>Woodcock</surname>, <given-names>R. W.</given-names></string-name> (<year>2002</year>). <article-title>Comparative longitudinal structural analyses of the growth and decline of multiple intellectual abilities over the life span</article-title>. <source><italic>Developmental Psychology</italic></source>, <volume>38</volume> (<issue>1</issue>), <fpage>115</fpage>–<lpage>142</lpage>. <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1037/0012-1649.38.1.115">https://doi.org/10.1037/0012-1649.38.1.115</ext-link></mixed-citation></ref>
<ref id="r28"><mixed-citation publication-type="journal"><string-name name-style="western"><surname>McLachlan</surname>, <given-names>G. J.</given-names></string-name>, <string-name name-style="western"><surname>Lee</surname>, <given-names>S. X.</given-names></string-name>, &amp; <string-name name-style="western"><surname>Rathnayake</surname>, <given-names>S. I.</given-names></string-name> (<year>2019</year>). <article-title>Finite Mixture Models</article-title>. <source><italic>Annual Review of Statistics and Its Application</italic></source>, <volume>6</volume> (<issue>1</issue>), <fpage>355</fpage>–<lpage>378</lpage>. <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1146/annurev-statistics-031017-100325">https://doi.org/10.1146/annurev-statistics-031017-100325</ext-link></mixed-citation></ref>
<ref id="r29"><mixed-citation publication-type="journal"><string-name name-style="western"><surname>Meredith</surname>, <given-names>W.</given-names></string-name> (<year>1993</year>). <article-title>Measurement invariance, factor analysis and factorial invariance</article-title>. <source><italic>Psychometrika</italic></source>, <volume>58</volume> (<issue>4</issue>), <fpage>525</fpage>–<lpage>543</lpage>. <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1007/BF02294825">https://doi.org/10.1007/BF02294825</ext-link></mixed-citation></ref>
<ref id="r30"><mixed-citation publication-type="journal"><string-name name-style="western"><surname>Nylund</surname>, <given-names>K. L.</given-names></string-name>, <string-name name-style="western"><surname>Asparouhov</surname>, <given-names>T.</given-names></string-name>, &amp; <string-name name-style="western"><surname>Muthén</surname>, <given-names>B. O.</given-names></string-name> (<year>2007</year>). <article-title>Deciding on the number of classes in Latent Class Analysis and Growth Mixture Modeling: A Monte Carlo simulation study</article-title>. <source><italic>Structural Equation Modeling: A Multidisciplinary Journal</italic></source>, <volume>14</volume> (<issue>4</issue>), <fpage>535</fpage>–<lpage>569</lpage>. <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1080/10705510701575396">https://doi.org/10.1080/10705510701575396</ext-link></mixed-citation></ref>
	
	<ref id="r30.5"><mixed-citation publication-type="data">Perez Alonzo, A. F. (2024). <italic>AndresFPA/ModelSelection_Simulation</italic> [GitHub project page containing code]. GitHub. <ext-link ext-link-type="uri" xlink:href="https://github.com/AndresFPA/ModelSelection_Simulation">https://github.com/AndresFPA/ModelSelection_Simulation</ext-link></mixed-citation></ref>
	
	<ref id="r30.75"><mixed-citation publication-type="data">Perez Alonzo, A. F. (2025). <italic>AndresFPA/mmgsem</italic> [GitHub project page containing code and documentation]. GitHub. <ext-link ext-link-type="uri" xlink:href="https://github.com/AndresFPA/ModelSelection_Simulation">https://github.com/AndresFPA/ModelSelection_Simulation</ext-link></mixed-citation></ref>
	
	<ref id="r31"><mixed-citation publication-type="book"><string-name name-style="western"><surname>Perez Alonso</surname>, <given-names>A. F.</given-names></string-name>, <string-name name-style="western"><surname>Rosseel</surname>, <given-names>Y.</given-names></string-name>, <string-name name-style="western"><surname>Vermunt</surname>, <given-names>J. K.</given-names></string-name>, &amp; <string-name name-style="western"><surname>De Roover</surname>, <given-names>K.</given-names></string-name> (<year>2024</year>). <article-title>Mixture multigroup structural equation modeling: A novel method for comparing structural relations across many groups</article-title>. <source><italic>Psychological Methods</italic></source>. <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1037/met0000667">https://doi.org/10.1037/met0000667</ext-link></mixed-citation></ref>
	
	<ref id="r31.5"><mixed-citation publication-type="data">
		
		<person-group person-group-type="author"><string-name name-style="western"><surname>Perez Alonso</surname>, <given-names>A. F.</given-names></string-name>, <string-name name-style="western"><surname>Vermunt</surname>, <given-names>J. K.</given-names></string-name>, <string-name name-style="western"><surname>Rosseel</surname>, <given-names>Y. D.</given-names></string-name>, &amp; <string-name name-style="western"><surname>De Roover</surname>, <given-names>K.</given-names></string-name></person-group>(<year>2025</year>). <italic>Supplementary materials to </italic>“Selecting the number of clusters in Mixture Multigroup Structural Equation Modeling” [Supplementary materials with simulation details and performance results]. PsychOpen GOLD. <pub-id pub-id-type="doi">10.23668/psycharchives.16173</pub-id></mixed-citation></ref>

	<ref id="r32"><mixed-citation publication-type="web"><collab>R Core Team</collab> (<year>2024</year>). <italic>R: A language and environment for statistical computing</italic>. R Project for Statistical Computing. <ext-link ext-link-type="uri" xlink:href="https://www.R-project.org/">https://www.R-project.org/</ext-link></mixed-citation></ref>
	<ref id="r33"><mixed-citation publication-type="journal"><string-name name-style="western"><surname>Rosseel</surname>, <given-names>Y.</given-names></string-name>, &amp; <string-name name-style="western"><surname>Loh</surname>, <given-names>W. W.</given-names></string-name> (<year>2022</year>). <article-title>A structural after measurement approach to structural equation modeling</article-title>. <source><italic>Psychological Methods</italic></source>,  <volume>29</volume>(<issue>3</issue>), <fpage>561</fpage>–<lpage>588</lpage>. <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1037/met0000503">https://doi.org/10.1037/met0000503</ext-link></mixed-citation></ref>
<ref id="r34"><mixed-citation publication-type="journal"><string-name name-style="western"><surname>Rutkowski</surname>, <given-names>L.</given-names></string-name>, &amp; <string-name name-style="western"><surname>Svetina</surname>, <given-names>D.</given-names></string-name> (<year>2014</year>). <article-title>Assessing the hypothesis of measurement invariance in the context of large-scale international surveys</article-title>. <source><italic>Educational and Psychological Measurement</italic></source>, <volume>74</volume> (<issue>1</issue>), <fpage>31</fpage>–<lpage>57</lpage>. <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1177/0013164413498257">https://doi.org/10.1177/0013164413498257</ext-link></mixed-citation></ref>
<ref id="r35"><mixed-citation publication-type="journal"><string-name name-style="western"><surname>Schwarz</surname>, <given-names>G.</given-names></string-name> (<year>1978</year>). <article-title>Estimating the dimension of a model</article-title>. <source><italic> Annals of Statistics</italic></source>, <volume>6</volume> (<issue>2</issue>), <fpage>461</fpage>–<lpage>464</lpage>. <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1214/aos/1176344136">https://doi.org/10.1214/aos/1176344136</ext-link></mixed-citation></ref>
<ref id="r36"><mixed-citation publication-type="journal"><string-name name-style="western"><surname>Steinley</surname>, <given-names>D.</given-names></string-name>, &amp; <string-name name-style="western"><surname>Brusco</surname>, <given-names>M. J.</given-names></string-name> (<year>2011</year>). <article-title>Evaluating mixture modeling for clustering: Recommendations and cautions</article-title>. <source><italic>Psychological Methods</italic></source>, <volume>16</volume> (<issue>1</issue>), <fpage>63</fpage>–<lpage>79</lpage>. <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1037/a0022673">https://doi.org/10.1037/a0022673</ext-link></mixed-citation></ref>
<ref id="r38"><mixed-citation publication-type="journal"><string-name name-style="western"><surname>Vandenberg</surname>, <given-names>R. J.</given-names></string-name>, &amp; <string-name name-style="western"><surname>Lance</surname>, <given-names>C. E.</given-names></string-name> (<year>2000</year>). <article-title>A review and synthesis of the measurement invariance literature: Suggestions, practices, and recommendations for organizational research</article-title>. <source><italic>Organizational Research Methods</italic></source>, <volume>3</volume> (<issue>1</issue>), <fpage>4</fpage>–<lpage>70</lpage>. <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1177/109442810031002">https://doi.org/10.1177/109442810031002</ext-link></mixed-citation></ref>
	<ref id="r37"><mixed-citation publication-type="journal"><string-name name-style="western"><surname>van den Bergh</surname>, <given-names>M.</given-names></string-name>, <string-name name-style="western"><surname>Schmittmann</surname>, <given-names>V. D.</given-names></string-name>, &amp; <string-name name-style="western"><surname>Vermunt</surname>, <given-names>J. K.</given-names></string-name> (<year>2017</year>). <article-title>Building Latent Class Trees, with an application to a study of social capital</article-title>. <source><italic>Methodology</italic></source>, <volume>13</volume> (<issue>Supplement 1</issue>), <fpage>13</fpage>–<lpage>22</lpage>. <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1027/1614-2241/a000128">https://doi.org/10.1027/1614-2241/a000128</ext-link></mixed-citation></ref>
<ref id="r39"><mixed-citation publication-type="book"><string-name name-style="western"><surname>Vermunt</surname>, <given-names>J. K.</given-names></string-name>, &amp; <string-name name-style="western"><surname>Magidson</surname>, <given-names>J.</given-names></string-name> (<year>2005</year>, <month>October</month>). <chapter-title>Structural Equation Modeling: Mixture Models</chapter-title>. In B. S. Everitt &amp; D. C. Howell (Eds.), <italic>Encyclopedia of statistics in behavioral science</italic>. <publisher-name>John Wiley &amp; Sons</publisher-name>. <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1002/0470013192.bsa600">https://doi.org/10.1002/0470013192.bsa600</ext-link></mixed-citation></ref>
<ref id="r40"><mixed-citation publication-type="book"><string-name name-style="western"><surname>Vermunt</surname>, <given-names>J. K.</given-names></string-name>, &amp; <string-name name-style="western"><surname>Magidson</surname>, <given-names>J.</given-names></string-name> (<year>2016</year>). <italic>Technical guide for Latent GOLD 5.1: Basic, advanced, and syntax</italic>. <publisher-name>Statistical Innovations</publisher-name>.</mixed-citation></ref>
<ref id="r41"><mixed-citation publication-type="journal"><string-name name-style="western"><surname>Vrieze</surname>, <given-names>S. I.</given-names></string-name> (<year>2012</year>). <article-title>Model selection and psychological theory: A discussion of the differences between the Akaike information criterion (AIC) and the Bayesian information criterion (BIC)</article-title>. <source><italic>Psychological Methods</italic></source>, <volume>17</volume> (<issue>2</issue>), <fpage>228</fpage>–<lpage>243</lpage>. <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1037/a0027127">https://doi.org/10.1037/a0027127</ext-link></mixed-citation></ref>
</ref-list>
			
	<sec sec-type="supplementary-material" id="sp1"><title>Supplementary Materials</title>
		<p>For this article, the following Supplementary Materials are available:
			<list list-type="bullet">
				<list-item><p>Code. (<xref ref-type="bibr" rid="r30.5">Perez Alonzo, 2024</xref>)</p></list-item>
				<list-item><p>Study materials. (<xref ref-type="bibr" rid="r31.5">Perez Alonso et al., 2025</xref>)</p></list-item>
			</list></p>
	</sec>
	<fn-group><fn fn-type="financial-disclosure">
		<p content-type="fn-title">This research was funded by a Vidi grant [VI.Vidi.201.133] awarded to Kim De Roover by the Netherlands Organization for Scientific Research (NWO).</p></fn></fn-group>
<fn-group>
<fn fn-type="conflict"><p>The authors have declared that no competing interests exist.</p></fn>
</fn-group>
<ack>
<p>The authors have no additional (i.e., non-financial) support to report.</p>
</ack>
</back>
</article>
