<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article
  PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD with MathML3 v1.2 20190208//EN" "JATS-journalpublishing1-mathml3.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xmlns:ali="http://www.niso.org/schemas/ali/1.0/" article-type="research-article" dtd-version="1.2" xml:lang="en">
<front>
<journal-meta><journal-id journal-id-type="publisher-id">METH</journal-id><journal-id journal-id-type="nlm-ta">Methodology</journal-id>
<journal-title-group>
<journal-title>Methodology</journal-title><abbrev-journal-title abbrev-type="pubmed">Methodology</abbrev-journal-title>
</journal-title-group>
<issn pub-type="ppub">1614-1881</issn>
<issn pub-type="epub">1614-2241</issn>
<publisher><publisher-name>PsychOpen</publisher-name></publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="publisher-id">meth.16411</article-id>
<article-id pub-id-type="doi">10.5964/meth.16411</article-id>
<article-categories>
<subj-group subj-group-type="heading"><subject>Original Article</subject></subj-group>
<subj-group subj-group-type="badge">
<subject>Code</subject>
<subject>Materials</subject>	
</subj-group>
</article-categories>
<title-group>
<article-title>Mixture Multigroup Bayesian SEM With Approximate Measurement Invariance for Comparing Structural Relations Across Many Groups</article-title>
<alt-title alt-title-type="right-running">Comparing Structural Relations Across Many Groups</alt-title>
<alt-title specific-use="APA-reference-style" xml:lang="en">Mixture multigroup Bayesian SEM with approximate measurement invariance for comparing structural relations across many groups</alt-title>
</title-group>
<contrib-group>
<contrib contrib-type="author" corresp="yes"><contrib-id contrib-id-type="orcid" authenticated="false">https://orcid.org/0009-0008-9372-4986</contrib-id><name name-style="western"><surname>Zhao</surname><given-names>Hongwei</given-names></name><xref ref-type="corresp" rid="cor1">*</xref><xref ref-type="aff" rid="aff1"><sup>1</sup></xref></contrib>
<contrib contrib-type="author"><contrib-id contrib-id-type="orcid" authenticated="false">https://orcid.org/0000-0001-9053-9330</contrib-id><name name-style="western"><surname>Vermunt</surname><given-names>Jeroen K.</given-names></name><xref ref-type="aff" rid="aff2"><sup>2</sup></xref></contrib>
<contrib contrib-type="author"><contrib-id contrib-id-type="orcid" authenticated="false">https://orcid.org/0000-0002-0299-0648</contrib-id><name name-style="western"><surname>De Roover</surname><given-names>Kim</given-names></name><xref ref-type="aff" rid="aff1"><sup>1</sup></xref><xref ref-type="aff" rid="aff2"><sup>2</sup></xref></contrib>
<contrib contrib-type="editor">
<name>
<surname>Aliri</surname>
<given-names>Jone</given-names>
</name>
<xref ref-type="aff" rid="aff3"/>
</contrib>
<aff id="aff1"><label>1</label><institution content-type="dept">Quantitative Psychology and Individual Differences</institution>, <institution>KU Leuven</institution>, <addr-line><city>Leuven</city></addr-line>, <country country="BE">Belgium</country></aff>
<aff id="aff2"><label>2</label><institution content-type="dept">Department of Methodology, Tilburg School of Social and Behavioral Sciences</institution>, <institution>Tilburg University</institution>, <addr-line><city>Tilburg</city></addr-line>, <country country="">the Netherlands</country></aff>
	<aff id="aff3">University of the Basque Country, Leioa, <country>Spain</country>.</aff>
</contrib-group>
<author-notes>
<corresp id="cor1"><label>*</label>Quantitative Psychology and Individual Differences, KU Leuven, Tiensestraat 102, 3000 Leuven, Belgium. <email xlink:href="hongwei.zhao@kuleuven.be">hongwei.zhao@kuleuven.be</email></corresp>
</author-notes>
<pub-date date-type="pub" publication-format="electronic"><day>18</day><month>12</month><year>2025</year></pub-date>
<pub-date pub-type="collection" publication-format="electronic"><year>2025</year></pub-date>
<volume>21</volume>
<issue>4</issue>

<fpage>286</fpage>
<lpage>312</lpage>
<history>
<date date-type="received">
<day>15</day>
<month>12</month>
<year>2024</year>
</date>
<date date-type="accepted">
<day>16</day>
<month>09</month>
<year>2025</year>
</date>
</history>
<permissions><copyright-year>2025</copyright-year><copyright-holder>Zhao, Vermunt, &amp; De Roover</copyright-holder><license license-type="open-access" specific-use="CC BY 4.0" xlink:href="https://creativecommons.org/licenses/by/4.0/"><ali:license_ref>https://creativecommons.org/licenses/by/4.0/</ali:license_ref><license-p>This is an open access article distributed under the terms of the Creative Commons Attribution 4.0 International License, CC BY 4.0, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.</license-p></license></permissions>
<abstract>
<p>In social sciences, researchers often compare relations between constructs, referred to as “structural relations”, across a large number of groups. This paper proposes Mixture Multigroup Bayesian SEM (MixMG-BSEM), a novel method for comparing structural relations across many groups while accounting for approximate measurement invariance in factor loadings. Traditional methods often assume exact measurement invariance, which may not reflect real-world data where small differences in measurement parameters commonly occur across many groups. MixMG-BSEM addresses this by using Multigroup Bayesian CFA with small-variance priors to allow for these small differences, and groups are then clustered based on their structural relations using Mixture Modeling. This is done in a stepwise estimation procedure built on the structural-after-measurement approach. By combining cluster-specific structural relations with small between-group differences in measurement parameters, MixMG-BSEM obtains a clustering that is driven only by the structural relations. The robustness and effectiveness of MixMG-BSEM are demonstrated through a simulation study.</p>
</abstract>
<kwd-group kwd-group-type="author"><kwd>Multigroup Bayesian SEM</kwd><kwd>Approximate Measurement Invariance</kwd><kwd>Mixture Modeling</kwd><kwd>Structural Relation</kwd></kwd-group>

</article-meta>
</front>
<body>
	<sec sec-type="intro" id="intro"><title/>	
<p>In social sciences, Structural Equation Modelling (SEM; <xref ref-type="bibr" rid="r5">Bollen, 1989</xref>; <xref ref-type="bibr" rid="r13">Hoyle, 2012</xref>) is widely used to investigate relations between constructs (e.g., emotions, motivation), referred to as “structural relations” within SEM. Researchers are often interested in how these structural relations vary across groups. For instance, <xref ref-type="bibr" rid="r24">Michael and Kyriakides (2023)</xref> examined how academic motivation mediated the effect of socioeconomic status on reading achievement among 15-year-old students and how this differed across 38 countries.</p>
<p>To study differences in structural relations, Multigroup SEM (MG-SEM) and Multilevel SEM (ML-SEM) can be used. MG-SEM estimates the structural relations for each group and allows testing whether they are equal across groups. ML-SEM captures variations in structural relations by normally distributed random effects around the overall mean estimate for each relation. Even though group-specific estimates of relations can be derived from random effects, only the mean and variance of each random effect are part of the model parameters, which makes ML-SEM more parsimonious, allowing for accurate parameter estimates in case of very small sample sizes per group. To pinpoint which groups have the same relations and for which groups they differ, MG-SEM and ML-SEM require pairwise comparisons of group-specific relations. As the number of groups increases, performing pairwise comparisons quickly becomes infeasible. For example, for 38 groups, this requires 703 pairwise comparisons per structural relation. To reduce the number of comparisons, mixture modeling (<xref ref-type="bibr" rid="r22">McLachlan &amp; Peel, 2000</xref>) can be used to cluster groups based on similarity of the structural relations. Before performing such a clustering, it is essential to ensure that the structural relations are validly comparable across groups and that they are the only source of differences driving the clustering.</p>
<p>In social sciences, the constructs of interest are typically unobserved or latent variables, also known as “factors” in SEM. SEM addresses their latent nature by including a measurement model (MM), which specifies how latent variables are measured by observed indicators (often questionnaire items), whereas the relations of interest among the latent variables are part of the structural model (SM). For valid comparisons of constructs and their relations, measurement invariance (MI) must hold across the groups. MI implies that the MM is equal across groups, meaning that the constructs are measured in the same way, so that observed differences reflect differences in the constructs rather than differences in measurement.</p>
<p>MI is examined at different levels by assessing the equality of different subsets of MM parameters. Configural invariance evaluates whether the factor structure is the same across groups, meaning that, in each group, the same set of indicators relates to a factor. The strength and direction of the relations between factors and indicators are quantified by factor loadings. Whereas configural invariance only deals with which factor loadings are non-zero, weak or metric invariance requires the loadings to be equal across groups. Next, strong and strict invariance impose equality of the items’ intercepts and residual or ‘unique’ variances, respectively. Metric invariance is a prerequisite for validly comparing structural relations, whereas strong and strict invariance are not required. When full metric invariance (i.e., invariance of all loadings) does not hold, partial metric invariance (i.e., invariance of some loadings) still enables valid comparisons of structural relations (<xref ref-type="bibr" rid="r6">Byrne et al., 1989</xref>), as long as the loading differences are captured in the model (e.g., by group specific loadings). The same holds for differences in item intercepts and unique variances.</p>
<p>When combining SEM with mixture modeling, groups can be clustered on their structural relations by making the structural relations cluster-specific (i.e., the same for all groups assigned to a cluster). In traditional mixture SEM methods (<xref ref-type="bibr" rid="r2">Arminger &amp; Stein, 1997</xref>; <xref ref-type="bibr" rid="r10">Dolan &amp; van der Maas, 1998</xref>; <xref ref-type="bibr" rid="r15">Jedidi et al., 1997</xref>), MM parameters can be specified as invariant or cluster-specific, implying that MM differences can either be ignored or captured by the same clustering. To cluster groups only on the structural relations rather than also on differences in measurement, a framework of novel mixture SEM methods emerged recently. <xref ref-type="bibr" rid="r29">Perez Alonso and colleagues (2024)</xref> introduced Mixture Multigroup SEM (MixMG-SEM), which combines MG-SEM with mixture modeling. <xref ref-type="bibr" rid="r43">Zhao and colleagues (2025a)</xref> proposed Mixture Multilevel SEM (MixML-SEM), which builds the mixture clustering onto the more parsimonious ML-SEM. The aim of both methods is to cluster groups specifically on the structural relations while accounting for measurement non-invariance, but the difference is that MixML-SEM uses Multilevel Confirmatory Factor Analysis (ML-CFA) with random effects to deal with MM differences, whereas MixMG-SEM uses Multigroup Confirmatory Factor Analysis (MG-CFA) with group-specific MM parameters. Their estimation builds on the stepwise “Structural-After-Measurement” (<xref ref-type="bibr" rid="r33">Rosseel &amp; Loh, 2024</xref>) approach, where the MM is estimated first, using either MG-CFA or ML-CFA, followed by the SM, which includes clustering the groups on their structural relations. For comparability of structural relations, both methods require at least partial metric invariance and impose exact equality for the invariant factor loadings (i.e., exact MI). However, with a large number of groups, achieving exact MI is often unrealistic. To address this, Multigroup Bayesian SEM (MG-BSEM; <xref ref-type="bibr" rid="r25">Muthén &amp; Asparouhov, 2012</xref>, <xref ref-type="bibr" rid="r26">2013a</xref>, <xref ref-type="bibr" rid="r27">2013b</xref>) with Approximate MI (AMI) uses priors with small variances for the MM parameters to allow for small differences across groups while keeping them approximately equal. In this paper, we present Mixture Multigroup BSEM (MixMG-BSEM), which extends MG-BSEM with mixture modeling to cluster groups on the structural relations while capturing approximate invariance of factor loadings.</p>
<p>MixMG-BSEM, MixMG-SEM and MixML-SEM differ in their first estimation step only, that is, in their MM and the corresponding MI assumptions. MixMG-SEM and MixML-SEM require exact invariance for (at least) some loadings, whereas the first step of MixMG-BSEM is a MG-CFA with Bayesian estimation (MG-BCFA) that assumes approximately invariant loadings. Approximate invariance lies between exact invariance (where parameters are exactly equal across groups) and non-invariance (where parameters can differ substantially across groups), where exact invariance is more closely approximated as the variances of the priors become smaller.</p>
<p>The paper is structured as follows: We begin with a description of MixMG-BSEM in the Method section. Next, we evaluate its performance through a Simulation Study. Finally, the Discussion section summarizes the main findings and addresses limitations and future directions.</p></sec>
<sec id="M1" sec-type="methods"><title>Method</title>
<p>As mentioned above, MixMG-BSEM is estimated in a stepwise manner, building on the SAM approach (see <xref ref-type="fig" rid="f1">Figure 1</xref>). In Step 1, MG-BCFA with small-variance priors is performed for each factor, and factor scores are extracted. In Step 2, these factor scores are used as single indicators to obtain group-specific factor covariances with Croon’s correction (<xref ref-type="bibr" rid="r8">Croon, 2002</xref>). In Step 3, the SM is estimated, including the clustering and the cluster-specific structural relations, using an Expectation-Maximization (<xref ref-type="bibr" rid="r9">Dempster et al., 1977</xref>) algorithm for maximum likelihood estimation. Note that Steps 2 and 3 are the same as for MixML-SEM and are therefore only briefly described below (for details, see <xref ref-type="bibr" rid="r43">Zhao et al., 2025a</xref>).</p><fig id="f1" position="anchor" fig-type="figure" orientation="portrait"><label>Figure 1</label><caption>
<title>Mixture Multigroup Bayesian SEM with Approximate Measurement Invariance</title></caption><graphic xlink:href="meth.16411-f1" position="anchor" orientation="portrait"/></fig>
<sec><title>Step 1: Measurement Model With Bayesian Approximate Measurement Invariance</title>
<p>The MM defines how the factors are measured by the items and MG-CFA is used to compare MMs across groups. Note that we estimate the MM per factor, which implies that we assume the factors to be independent in Step 1. Indicating an individual in Group <inline-formula><mml:math id="m1"><mml:mi>g</mml:mi></mml:math></inline-formula> (<inline-formula><mml:math id="m2"><mml:mi>g</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mo>…</mml:mo><mml:mo>,</mml:mo><mml:mi>G</mml:mi><mml:mo>)</mml:mo></mml:math></inline-formula> by <inline-formula><mml:math id="m3"><mml:msub><mml:mrow><mml:mi>n</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> and gathering the responses on the <inline-formula><mml:math id="m4"><mml:msub><mml:mrow><mml:mi>J</mml:mi></mml:mrow><mml:mrow><mml:mi>q</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> items measuring Factor <inline-formula><mml:math id="m5"><mml:mi>q</mml:mi></mml:math></inline-formula> (<inline-formula><mml:math id="m6"><mml:mi>q</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mo>…</mml:mo><mml:mo>,</mml:mo><mml:mi>Q</mml:mi></mml:math></inline-formula>) in the Vector <inline-formula><mml:math id="m7"><mml:msub><mml:mrow><mml:mi mathvariant="bold">x</mml:mi></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>n</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:msub></mml:math></inline-formula>, the MG-CFA model for Factor <italic>q</italic> is expressed as:</p><disp-formula id="e1"><label>1</label><mml:math id="mml-eqn-1" display="block"><mml:mtable><mml:mtr><mml:mtd><mml:mrow><mml:maligngroup/><mml:msub><mml:mrow><mml:mi mathvariant="bold">x</mml:mi></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>n</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:msub><mml:mrow><mml:mi mathvariant="bold">τ</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub><mml:mo>+</mml:mo><mml:msub><mml:mrow><mml:mi mathvariant="bold">λ</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub><mml:msub><mml:mrow><mml:mi>η</mml:mi></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>n</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:msub><mml:mo>+</mml:mo><mml:msub><mml:mrow><mml:mi mathvariant="bold">ϵ</mml:mi></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>n</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:msub><mml:mi> </mml:mi><mml:mi>w</mml:mi><mml:mi>i</mml:mi><mml:mi>t</mml:mi><mml:mi>h</mml:mi><mml:mi> </mml:mi><mml:msub><mml:mrow><mml:mi mathvariant="bold">ϵ</mml:mi></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>n</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:msub><mml:mo>~</mml:mo><mml:mi> </mml:mi><mml:mi>M</mml:mi><mml:mi>V</mml:mi><mml:mi>N</mml:mi><mml:mfenced separators="|"><mml:mrow><mml:mn>0</mml:mn><mml:mo>,</mml:mo><mml:msub><mml:mrow><mml:mi mathvariant="bold">Θ</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mfenced></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>where <inline-formula><mml:math id="m9"><mml:msub><mml:mrow><mml:mi mathvariant="bold">τ</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> is a<inline-formula><mml:math id="m10"><mml:mi> </mml:mi><mml:msub><mml:mrow><mml:mi>J</mml:mi></mml:mrow><mml:mrow><mml:mi>q</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>-dimensional vector of intercepts for Group <inline-formula><mml:math id="m11"><mml:mi>g</mml:mi></mml:math></inline-formula>, <inline-formula><mml:math id="m12"><mml:msub><mml:mrow><mml:mi mathvariant="bold">λ</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> is a <inline-formula><mml:math id="m13"><mml:msub><mml:mrow><mml:mi>J</mml:mi></mml:mrow><mml:mrow><mml:mi>q</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>-dimensional vector of factor loadings (i.e., item-factor relations) for Group <inline-formula><mml:math id="m14"><mml:mi>g</mml:mi></mml:math></inline-formula>, <inline-formula><mml:math id="m15"><mml:msub><mml:mrow><mml:mi>η</mml:mi></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>n</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:msub></mml:math></inline-formula> denotes the latent variable score for the individual, and <inline-formula><mml:math id="m16"><mml:msub><mml:mrow><mml:mi mathvariant="bold">ϵ</mml:mi></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>n</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:msub></mml:math></inline-formula> is a <inline-formula><mml:math id="m17"><mml:msub><mml:mrow><mml:mi>J</mml:mi></mml:mrow><mml:mrow><mml:mi>q</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>-dimensional vector of residuals, with the diagonal of <inline-formula><mml:math id="m18"><mml:msub><mml:mrow><mml:mi mathvariant="bold">Θ</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> containing the group-specific unique variances of the items. To set the scale of each factor, one can either set its variance to one or use the marker variable approach by fixing one loading (ideally, a strong and invariant loading) to one, for each group. In this paper, we adopt the marker variable approach to ensure that a one-unit change in the underlying factor has the same meaning across groups.</p>
<p>Since small differences in MM parameters are common across many groups and still allow for latent variable comparisons, we apply the assumption of approximate metric invariance (i.e., <inline-formula><mml:math id="m19"><mml:msub><mml:mrow><mml:mi mathvariant="bold">λ</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub><mml:mo>≈</mml:mo><mml:mi mathvariant="bold">λ</mml:mi></mml:math></inline-formula> for all Groups <inline-formula><mml:math id="m20"><mml:mi>g</mml:mi></mml:math></inline-formula>) instead of exact invariance (i.e., <inline-formula><mml:math id="m21"><mml:msub><mml:mrow><mml:mi mathvariant="bold">λ</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mi mathvariant="bold">λ</mml:mi></mml:math></inline-formula>) in Step 1 of MixMG-BSEM. This is accomplished by using MG-CFA with Bayesian estimation<xref ref-type="fn" rid="fn1"><sup>1</sup></xref><fn id="fn1"><label>1</label>
<p>The possibility to use a different estimator in each step, such as Bayesian estimation for the MM and maximum likelihood for the SM (see also <xref ref-type="bibr" rid="r43">Zhao et al., 2025a</xref>) is an important advantage of the SAM approach.</p></fn> (MG-BCFA) and applying small-variance, normally distributed priors to the corresponding factor loadings, which constrain the group-specific loadings to be approximately equal (<xref ref-type="bibr" rid="r25">Muthén &amp; Asparouhov, 2012</xref>). For this, both <italic>Mplus</italic> (<xref ref-type="bibr" rid="r28">Muthén &amp; Muthén, 1998</xref>) and the R-package <italic>blavaan</italic> (<xref ref-type="bibr" rid="r23">Merkle et al., 2021</xref>) are available, but we use <italic>blavaan</italic> by default because it is free and open-source. In <italic>blavaan</italic>, AMI is achieved by applying small-variance priors in every group except for the reference group, which is the first group (by default). A non-informative prior is used for the parameter in the first group and the parameter estimate for this group is used as the mean of the small-variance priors for that same parameter in the other groups.</p>
<p>Since Bayesian estimation can be computationally challenging, two measures are taken to lower the computation time of Step 1 of MixMG-BSEM: (1) the data are centered per group to remove the mean structure (i.e., <inline-formula><mml:math id="m22"><mml:msub><mml:mrow><mml:mi mathvariant="bold">τ</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mn>0</mml:mn></mml:math></inline-formula> and <inline-formula><mml:math id="m23"><mml:msub><mml:mrow><mml:mi mathvariant="bold">α</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mn>0</mml:mn></mml:math></inline-formula>), which is irrelevant to the comparison of structural relations, and (2) MG-BCFA is performed for each factor separately, which is in line with the “measurement blocks” approach in SAM (<xref ref-type="bibr" rid="r33">Rosseel &amp; Loh, 2024</xref>) with one factor per measurement block. This approach lowers the number of parameters to be estimated and also enhances the model's robustness against MM misspecifications, such as a few unmodeled crossloadings.</p>
<p>In this paper, we assume that all factor loadings, except for the marker variable loadings, are approximately invariant, while the unique variances and factor variances are estimated as group-specific parameters (i.e., with the default non-informative priors per group). Note that it is harmless to specify exactly invariant loadings as approximately invariant since they will then be estimated as nearly identical across groups. Of course, in practice, combinations of exactly and approximately invariant loadings can be applied in Step 1 of MixMG-BSEM. Moreover, in theory, all combinations of exact invariance, approximate invariance and non-invariance can be used, but complex combinations may cause convergence problems.</p>
<p>To determine which parameters are (approximately) invariant or non-invariant, MI testing should be performed prior to using MixMG-BSEM. Note that, if exact invariance does not hold for a parameter, standard MG-CFA requires a tedious process of comparing group-specific parameter estimates to determine whether differences reflect non-invariance or approximate invariance. Instead, MG-BCFA allows to test the tenability of AMI directly by imposing small-variance priors on MM parameters and assessing model fit. <xref ref-type="bibr" rid="r25">Muthén and Asparouhov (2012)</xref> recommend starting with a very small variance (e.g., 0.001) and, if needed, the priors’ variances can be increased to reach a good model fit. In this way, MG-BCFA provides information on how large the parameter differences are (i.e., on the level of AMI). Model fit can be assessed using the posterior predictive <italic>p</italic> value (<xref ref-type="bibr" rid="r11">Gelman et al., 1996</xref>), but it is not very sensitive to the prior variances in case of large samples. Other fit measures include the Bayesian RMSEA (BRMSEA; <xref ref-type="bibr" rid="r12">Hoofs et al., 2018</xref>) and the Deviance Information Criterion (DIC; <xref ref-type="bibr" rid="r35">Spiegelhalter et al., 2002</xref>). The DIC balances model fit (i.e., the posterior mean deviance) and complexity (i.e., the effective number of parameters) in Bayesian models, with smaller values indicating a better balance. Regarding the prior selection in MG-BSEM, <xref ref-type="bibr" rid="r16">Kim et al. (2017)</xref> found that the DIC often selected models with smaller prior variances when the sample size was small and <xref ref-type="bibr" rid="r31">Pokropek et al. (2020)</xref> found that the DIC performed better as sample size increased, and recommended using the DIC with thresholds tailored to different sample sizes.</p>
<p>Once the marker variables and the approximately invariant loadings are confirmed by the MI testing, we obtain the specification of the MG-BCFA model that corresponds to the first step of MixMG-BSEM. In the next step, we need estimates of the factor scores and their uncertainty. To this end, the means and standard deviations of the posterior distributions of the individuals’ factor scores (i.e., estimated latent variable scores) are appended to the data file.</p></sec>
<sec><title>Step 2: Single-Indicator Approach to Obtain Group-Specific Factor Covariances</title>
<p>In a single-indicator approach, the factor scores are used as the “observed” proxy (or a single indicator) for the latent variable (<xref ref-type="bibr" rid="r38">Vermunt, 2025</xref>). Since factor scores are only estimates of the true latent variable scores, we apply Croon’s correction (2002) to the factor score covariances (<inline-formula><mml:math id="m24"><mml:mi mathvariant="normal">c</mml:mi><mml:mi mathvariant="normal">o</mml:mi><mml:mi mathvariant="normal">v</mml:mi><mml:mfenced separators="|"><mml:mrow><mml:msub><mml:mrow><mml:mi mathvariant="bold">f</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mfenced></mml:math></inline-formula>) to obtain unbiased estimates of the true latent variable covariances (<inline-formula><mml:math id="m25"><mml:mi mathvariant="normal">c</mml:mi><mml:mi mathvariant="normal">o</mml:mi><mml:mi mathvariant="normal">v</mml:mi><mml:mfenced separators="|"><mml:mrow><mml:msub><mml:mrow><mml:mi mathvariant="bold">η</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mfenced></mml:math></inline-formula>), here denoted as <inline-formula><mml:math id="m26"><mml:msubsup><mml:mrow><mml:mi mathvariant="bold">Φ</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow><mml:mrow><mml:mi>s</mml:mi><mml:mn>2</mml:mn></mml:mrow></mml:msubsup></mml:math></inline-formula>:</p>
	<disp-formula id="e___1"><label>2</label><mml:math id="m27" display="block"><mml:mtable><mml:mtr><mml:mtd><mml:mrow><mml:maligngroup/><mml:msubsup><mml:mrow><mml:mi mathvariant="bold">Φ</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow><mml:mrow><mml:mi>s</mml:mi><mml:mn>2</mml:mn></mml:mrow></mml:msubsup><mml:mo>=</mml:mo><mml:mi> </mml:mi><mml:msup><mml:mrow><mml:mover accent="true"><mml:mrow><mml:msub><mml:mrow><mml:mi mathvariant="bold">Λ</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo>^</mml:mo></mml:mover><mml:mi mathvariant="normal"> </mml:mi></mml:mrow><mml:mrow><mml:mo>-</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msup><mml:mfenced separators="|"><mml:mrow><mml:mi mathvariant="normal">c</mml:mi><mml:mi mathvariant="normal">o</mml:mi><mml:mi mathvariant="normal">v</mml:mi><mml:mfenced separators="|"><mml:mrow><mml:msub><mml:mrow><mml:mi mathvariant="bold">F</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mfenced><mml:mo>-</mml:mo><mml:mover accent="true"><mml:mrow><mml:msub><mml:mrow><mml:mi mathvariant="bold">Θ</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow></mml:mfenced><mml:msup><mml:mrow><mml:mfenced separators="|"><mml:mrow><mml:mover accent="true"><mml:mrow><mml:msub><mml:mrow><mml:mi mathvariant="bold">Λ</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo>^</mml:mo></mml:mover><mml:mi mathvariant="normal"> </mml:mi><mml:mi mathvariant="normal">'</mml:mi></mml:mrow></mml:mfenced></mml:mrow><mml:mrow><mml:mo>-</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msup></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>where <inline-formula><mml:math id="m28"><mml:mover accent="true"><mml:mrow><mml:msub><mml:mrow><mml:mi mathvariant="bold">Λ</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:math></inline-formula> corresponds to the <inline-formula><mml:math id="m29"><mml:mi>Q</mml:mi><mml:mo>×</mml:mo><mml:mi>Q</mml:mi></mml:math></inline-formula> diagonal matrix of group-specific factor loadings (reflecting the reliability of the factor scores) and <inline-formula><mml:math id="m30"><mml:mover accent="true"><mml:mrow><mml:msub><mml:mrow><mml:mi mathvariant="bold">Θ</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:math></inline-formula> is the <inline-formula><mml:math id="m31"><mml:mi>Q</mml:mi><mml:mo>×</mml:mo><mml:mi>Q</mml:mi></mml:math></inline-formula> diagonal matrix of group-specific unique variances. These estimates correspond to the MM parameters of the single indicators of the factors (i.e., the factor scores) rather than the original, observed indicators. These MM parameters are derived from the posterior mean and standard deviation estimates for the factor scores, obtained from Step 1. For details, please refer to Equations (7–8) in the MixML-SEM paper (<xref ref-type="bibr" rid="r43">Zhao et al., 2025a</xref>).</p></sec>
<sec><title>Step 3: Structural Model With Mixture Clustering of the Groups</title>
<p>In Step 3, MixMG-BSEM clusters the groups and estimates cluster-specific structural relations. The SM is thus conditional on the cluster membership, <inline-formula><mml:math id="m32"><mml:msub><mml:mrow><mml:mi>z</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, which denotes whether Group <inline-formula><mml:math id="m33"><mml:mi>g</mml:mi></mml:math></inline-formula> belongs to Cluster <inline-formula><mml:math id="m34"><mml:mi>k</mml:mi></mml:math></inline-formula>. Whereas the true cluster membership is assumed to be either 1 or 0, its estimation, <inline-formula><mml:math id="m35"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>z</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>g</mml:mi><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, ranges from 0 to 1 and represents the probability of Group <inline-formula><mml:math id="m36"><mml:mi>g</mml:mi><mml:mi> </mml:mi></mml:math></inline-formula>belonging to Cluster <inline-formula><mml:math id="m37"><mml:mi>k</mml:mi></mml:math></inline-formula>. The model-implied factor covariance matrix <inline-formula><mml:math id="m38"><mml:msub><mml:mrow><mml:mi mathvariant="bold">Φ</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, given that <inline-formula><mml:math id="m39"><mml:msub><mml:mrow><mml:mi>z</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:math></inline-formula>, is defined as:</p><disp-formula id="eqn-3"><label>3</label><mml:math id="mm40" display="block"><mml:mtable><mml:mtr><mml:mtd><mml:mrow><mml:maligngroup/><mml:msub><mml:mrow><mml:mi mathvariant="bold">Φ</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:msup><mml:mrow><mml:mfenced separators="|"><mml:mrow><mml:mi mathvariant="bold">Ι</mml:mi><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mi mathvariant="bold">Β</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mfenced></mml:mrow><mml:mrow><mml:mo>-</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msup><mml:msub><mml:mrow><mml:mi mathvariant="bold">Ψ</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:msup><mml:mrow><mml:mfenced separators="|"><mml:mrow><mml:mi mathvariant="bold">Ι</mml:mi><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mi mathvariant="bold">Β</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mfenced></mml:mrow><mml:mrow><mml:mo>-</mml:mo><mml:msup><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi mathvariant="normal">'</mml:mi></mml:mrow></mml:msup></mml:mrow></mml:msup></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>where <inline-formula><mml:math id="m41"><mml:msub><mml:mrow><mml:mi mathvariant="bold">Β</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> contains the cluster-specific regression coefficients between latent variables, and <inline-formula><mml:math id="m42"><mml:msub><mml:mrow><mml:mi mathvariant="bold">Ψ</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> is the residual factor covariance matrix, which is specified as group-and-cluster-specific to ensure that clustering is driven only by the regressions <inline-formula><mml:math id="m43"><mml:msub><mml:mrow><mml:mi mathvariant="bold">Β</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> (for details, see <xref ref-type="bibr" rid="r29">Perez Alonso et al., 2024</xref>). The SM is estimated with maximum likelihood estimation using <inline-formula><mml:math id="m44"><mml:msubsup><mml:mrow><mml:mi mathvariant="bold">Φ</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow><mml:mrow><mml:mi>s</mml:mi><mml:mn>2</mml:mn></mml:mrow></mml:msubsup></mml:math></inline-formula> as input.</p>
<p>For the mixture clustering in MixMG-BSEM, it is assumed that the (true) latent variable scores <inline-formula><mml:math id="m45"><mml:msub><mml:mrow><mml:mi mathvariant="bold">η</mml:mi></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>n</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:msub></mml:math></inline-formula> are sampled from a mixture of <inline-formula><mml:math id="m46"><mml:mi>K</mml:mi></mml:math></inline-formula> multivariate normal distributions. Specifically, all latent variable scores of Group <inline-formula><mml:math id="m47"><mml:mi>g</mml:mi></mml:math></inline-formula>, <inline-formula><mml:math id="m48"><mml:msub><mml:mrow><mml:mi mathvariant="bold">H</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, are assumed to be sampled from the same distribution:</p><disp-formula id="eqn-4"><label>4</label><mml:math id="m49" display="block"><mml:mtable><mml:mtr><mml:mtd><mml:mrow><mml:maligngroup/><mml:mi>f</mml:mi><mml:mfenced separators="|"><mml:mrow><mml:msub><mml:mrow><mml:mi mathvariant="bold">H</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub><mml:mo>;</mml:mo><mml:mi>υ</mml:mi></mml:mrow></mml:mfenced><mml:mo>=</mml:mo><mml:mrow><mml:munderover><mml:mo stretchy="false">∑</mml:mo><mml:mrow><mml:mi>k</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>K</mml:mi></mml:mrow></mml:munderover><mml:mrow><mml:msub><mml:mrow><mml:mi mathvariant="normal">π</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">k</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mrow><mml:mi> </mml:mi><mml:mrow><mml:munderover><mml:mo stretchy="false">∏</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>n</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>N</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:munderover><mml:mrow><mml:mi>M</mml:mi><mml:mi>V</mml:mi><mml:mi>N</mml:mi><mml:mfenced separators="|"><mml:mrow><mml:msub><mml:mrow><mml:mi mathvariant="bold">η</mml:mi></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>n</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:msub><mml:mo>;</mml:mo><mml:msub><mml:mrow><mml:mi mathvariant="bold">α</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub><mml:msub><mml:mrow><mml:mo>,</mml:mo><mml:mi mathvariant="bold">Φ</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mfenced></mml:mrow></mml:mrow><mml:mi> </mml:mi><mml:mi>w</mml:mi><mml:mi>i</mml:mi><mml:mi>t</mml:mi><mml:mi>h</mml:mi><mml:mi> </mml:mi><mml:mrow><mml:munderover><mml:mo stretchy="false">∑</mml:mo><mml:mrow><mml:mi>k</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>K</mml:mi></mml:mrow></mml:munderover><mml:mrow><mml:msub><mml:mrow><mml:mi mathvariant="normal">π</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">k</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mrow><mml:mo>=</mml:mo><mml:mi> </mml:mi><mml:mn>1</mml:mn></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
	<p>where <inline-formula><mml:math id="m50"><mml:mi>f</mml:mi></mml:math></inline-formula> is the population density function, <inline-formula><mml:math id="m51"><mml:mi>υ</mml:mi></mml:math></inline-formula> represents the set of population parameters, and <inline-formula><mml:math id="m52"><mml:msub><mml:mrow><mml:mi mathvariant="normal">π</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> is the prior probability that Group <inline-formula><mml:math id="m53"><mml:mi>g</mml:mi></mml:math></inline-formula> belongs to Cluster <inline-formula><mml:math id="m54"><mml:mi>k</mml:mi></mml:math></inline-formula>. The scores in <inline-formula><mml:math id="m55"><mml:msub><mml:mrow><mml:mi mathvariant="bold">H</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> are assumed to follow a normal distribution with <inline-formula><mml:math id="m56"><mml:msub><mml:mrow><mml:mi mathvariant="bold">α</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> as the factor mean (which is zero due to centering) and <inline-formula><mml:math id="m57"><mml:msub><mml:mrow><mml:mi mathvariant="bold">Φ</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> as the factor covariance matrix. The unknown parameters <inline-formula><mml:math id="m58"><mml:mi>υ</mml:mi></mml:math></inline-formula> are estimated by maximizing the following log-likelihood function:</p><disp-formula id="eqn-5"><label>5</label><mml:math id="m59" display="block"><mml:mtable><mml:mtr><mml:mtd><mml:mrow><mml:maligngroup/><mml:mrow><mml:mrow><mml:mi mathvariant="normal">log</mml:mi></mml:mrow><mml:mo>⁡</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>L</mml:mi></mml:mrow><mml:mrow><mml:mi>η</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mrow><mml:mo>=</mml:mo><mml:mrow><mml:mrow><mml:mi mathvariant="normal">log</mml:mi></mml:mrow><mml:mo>⁡</mml:mo><mml:mrow><mml:mfenced separators="|"><mml:mrow><mml:mrow><mml:munderover><mml:mo stretchy="false">∏</mml:mo><mml:mrow><mml:mi>g</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>G</mml:mi></mml:mrow></mml:munderover><mml:mrow><mml:mrow><mml:munderover><mml:mo stretchy="false">∑</mml:mo><mml:mrow><mml:mi>k</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>K</mml:mi></mml:mrow></mml:munderover><mml:mrow><mml:msub><mml:mrow><mml:mi>π</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:mfrac><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:msup><mml:mrow><mml:mfenced separators="|"><mml:mrow><mml:mn>2</mml:mn><mml:mi>π</mml:mi></mml:mrow></mml:mfenced></mml:mrow><mml:mrow><mml:mrow><mml:mrow><mml:mi>Q</mml:mi></mml:mrow><mml:mo>/</mml:mo><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:mrow></mml:mrow></mml:msup><mml:mo>|</mml:mo><mml:msup><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mrow><mml:mi mathvariant="bold">Φ</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:mo>)</mml:mo><mml:mo>|</mml:mo></mml:mrow><mml:mrow><mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mo>/</mml:mo><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:mrow></mml:mrow></mml:msup></mml:mrow></mml:mfrac><mml:msup><mml:mrow><mml:mrow><mml:mrow><mml:mi mathvariant="normal">exp</mml:mi></mml:mrow><mml:mo>⁡</mml:mo><mml:mrow><mml:mfenced separators="|"><mml:mrow><mml:mo>-</mml:mo><mml:mfrac><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:mfrac><mml:mi>t</mml:mi><mml:mi>r</mml:mi><mml:mfenced separators="|"><mml:mrow><mml:msubsup><mml:mrow><mml:mi mathvariant="bold">Φ</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow><mml:mrow><mml:mi>s</mml:mi><mml:mn>2</mml:mn></mml:mrow></mml:msubsup><mml:msup><mml:mrow><mml:msub><mml:mrow><mml:mi mathvariant="bold">Φ</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mo>-</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msup></mml:mrow></mml:mfenced></mml:mrow></mml:mfenced></mml:mrow></mml:mrow></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>N</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:msup></mml:mrow></mml:mrow></mml:mrow></mml:mrow></mml:mrow></mml:mfenced></mml:mrow></mml:mrow></mml:mrow></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mrow><mml:maligngroup/><mml:mo>=</mml:mo><mml:mi> </mml:mi><mml:mrow><mml:mrow><mml:mrow><mml:munderover><mml:mo stretchy="false">∑</mml:mo><mml:mrow><mml:mi>g</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>G</mml:mi></mml:mrow></mml:munderover><mml:mrow><mml:mi mathvariant="normal">log</mml:mi></mml:mrow></mml:mrow></mml:mrow><mml:mo>⁡</mml:mo><mml:mrow><mml:mfenced separators="|"><mml:mrow><mml:mrow><mml:munderover><mml:mo stretchy="false">∑</mml:mo><mml:mrow><mml:mi>k</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>K</mml:mi></mml:mrow></mml:munderover><mml:mrow><mml:msub><mml:mrow><mml:mi>π</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:mfrac><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:msup><mml:mrow><mml:mfenced separators="|"><mml:mrow><mml:mn>2</mml:mn><mml:mi>π</mml:mi></mml:mrow></mml:mfenced></mml:mrow><mml:mrow><mml:mrow><mml:mrow><mml:mi>Q</mml:mi></mml:mrow><mml:mo>/</mml:mo><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:mrow></mml:mrow></mml:msup><mml:mo>|</mml:mo><mml:msup><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mrow><mml:mi mathvariant="bold">Φ</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:mo>)</mml:mo><mml:mo>|</mml:mo></mml:mrow><mml:mrow><mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mo>/</mml:mo><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:mrow></mml:mrow></mml:msup></mml:mrow></mml:mfrac><mml:msup><mml:mrow><mml:mrow><mml:mrow><mml:mi mathvariant="normal">exp</mml:mi></mml:mrow><mml:mo>⁡</mml:mo><mml:mrow><mml:mfenced separators="|"><mml:mrow><mml:mo>-</mml:mo><mml:mfrac><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:mfrac><mml:mi>t</mml:mi><mml:mi>r</mml:mi><mml:mfenced separators="|"><mml:mrow><mml:msubsup><mml:mrow><mml:mi mathvariant="bold">Φ</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow><mml:mrow><mml:mi>s</mml:mi><mml:mn>2</mml:mn></mml:mrow></mml:msubsup><mml:msup><mml:mrow><mml:msub><mml:mrow><mml:mi mathvariant="bold">Φ</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mo>-</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msup></mml:mrow></mml:mfenced></mml:mrow></mml:mfenced></mml:mrow></mml:mrow></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>N</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:msup></mml:mrow></mml:mrow></mml:mrow></mml:mfenced></mml:mrow></mml:mrow></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>where <inline-formula><mml:math id="m60"><mml:msubsup><mml:mrow><mml:mi mathvariant="bold">Φ</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow><mml:mrow><mml:mi>s</mml:mi><mml:mn>2</mml:mn></mml:mrow></mml:msubsup></mml:math></inline-formula> is the group-specific factor covariance matrix from Step 2 (Equation 2), and <inline-formula><mml:math id="m61"><mml:msub><mml:mrow><mml:mi mathvariant="bold">Φ</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> is the group-and-cluster-specific factor covariance matrix from Step 3 (Equation 3). The maximum likelihood estimation is performed using the EM algorithm (<xref ref-type="bibr" rid="r9">Dempster et al., 1977</xref>). Specifically, in the E-step, the algorithm estimates the classification probabilities <inline-formula><mml:math id="m62"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>z</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>g</mml:mi><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> given the current parameter estimates. In the M-step, the algorithm estimates the unknown parameters <inline-formula><mml:math id="m51.5"><mml:mi>𝜐</mml:mi></mml:math></inline-formula> given the classification probabilities obtained from the E-step. The E- and M-steps are iterated until convergence. A multi-start procedure is applied to mitigate convergence to local maxima, where the converged solution with the highest loglikelihood across the different starts is selected as the final result. For an in-depth explanation of the technical details of Step 3, readers are referred to Appendix A of the paper by <xref ref-type="bibr" rid="r29">Perez Alonso et al. (2024)</xref>.</p></sec></sec>
<sec sec-type="other1"><title>Simulation</title>
<p>In the simulation study, we evaluated the performance of MixMG-BSEM, assuming the true number of clusters was known. Firstly, we aimed to examine how MixMG-BSEM’s performance was affected by the within-group sample size, the number of groups, the number of clusters, the item reliability, the AMI of the loadings, and the size of (differences in) regression parameters. On top of that, since the first step of MixMG-BSEM estimates the MM per factor, we evaluated the consequences of ignoring crossloadings in this step. Literature on traditional SEM has shown that factor correlations tend to be overestimated when crossloadings are constrained to zero (e.g., <xref ref-type="bibr" rid="r4">Asparouhov et al., 2015</xref>; <xref ref-type="bibr" rid="r21">Marsh et al., 2009</xref>, <xref ref-type="bibr" rid="r19">2010</xref>, <xref ref-type="bibr" rid="r20">2014</xref>), which may affect the comparison of structural relations. However, given its stepwise estimation and measurement block approach, MixMG-BSEM may be relatively robust to overlooked crossloadings (<xref ref-type="bibr" rid="r33">Rosseel &amp; Loh, 2024</xref>), but the recovery of clusters and regression parameters may still decline in case of multiple crossloadings. Secondly, in terms of the analysis, we examined the impact of a key aspect of the Bayesian estimation; that is, the impact of different prior variances for the loadings on the recovery of clusters and cluster-specific regressions. We expected that using too narrow priors might fail to capture the loading differences across groups, which may affect the estimation of and clustering on the structural relations. Additionally, we also evaluated which prior was selected by the Deviance Information Criterion (DIC), since selecting this prior is an important step in empirical practice.</p>
<p>In a complete factorial design, the following factors were manipulated:</p>
<list id="L1" list-type="order">
<list-item>
<p>Total number of groups <inline-formula><mml:math id="m63"><mml:mi>G</mml:mi></mml:math></inline-formula> (2 levels): 24, 48.</p></list-item>
<list-item>
<p>Within-group sample size <inline-formula><mml:math id="m64"><mml:msub><mml:mrow><mml:mi>N</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> (3 levels): 50, 100, 200.</p></list-item>
<list-item>
<p>Number of clusters <inline-formula><mml:math id="m65"><mml:mi>K</mml:mi></mml:math></inline-formula> (2 levels): 2, 4.</p></list-item>
<list-item>
<p>Size of regression parameters <inline-formula><mml:math id="m66"><mml:mi>β</mml:mi></mml:math></inline-formula> (3 levels): 0.2, 0.3, 0.4.</p></list-item>
<list-item>
<p>Item reliability (2 levels): high, low.</p></list-item>
<list-item>
<p>Level of AMI for loadings (5 levels): 0.001, 0.005, 0.01, 0.05, 0.1.</p></list-item>
<list-item>
<p>Size of crossloadings (3 levels): 0, 0.2, 0.4.</p></list-item>
</list>
<p>We chose a minimum of 24 groups with group sizes <inline-formula><mml:math id="m67"><mml:msub><mml:mrow><mml:mi>N</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> ranging from 50 to 200, which partially correspond to the group sizes in other simulation studies on Bayesian AMI (<xref ref-type="bibr" rid="r16">Kim et al., 2017</xref>; <xref ref-type="bibr" rid="r18">Lek et al., 2018</xref>). The number of groups in each cluster depended on the number of groups <inline-formula><mml:math id="m68"><mml:mi>G</mml:mi></mml:math></inline-formula>, and the number of Clusters <inline-formula><mml:math id="m69"><mml:mi>K</mml:mi></mml:math></inline-formula>, where each cluster contained an equal number of groups. Note that larger <inline-formula><mml:math id="m70"><mml:mi>G</mml:mi></mml:math></inline-formula>, larger <inline-formula><mml:math id="m71"><mml:msub><mml:mrow><mml:mi>N</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, and smaller <inline-formula><mml:math id="m72"><mml:mi>K</mml:mi></mml:math></inline-formula> result in larger within-cluster sample sizes, which were expected to improve the performance of MixMG-BSEM.</p>
<p>The data were generated from a SEM model with four latent variables, each measured by five items (see <xref ref-type="fig" rid="f2">Figure 2</xref>), as in <xref ref-type="bibr" rid="r29">Perez Alonso et al. (2024)</xref> and <xref ref-type="bibr" rid="r43">Zhao et al. (2025a)</xref>. Specifically, the data were generated from a multivariate normal distribution (MVN) with covariance matrix <inline-formula><mml:math id="m73"><mml:msub><mml:mrow><mml:mi mathvariant="bold">Σ</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, determined by the parameters <inline-formula><mml:math id="m74"><mml:msub><mml:mrow><mml:mi mathvariant="bold">Β</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, <inline-formula><mml:math id="m75"><mml:msub><mml:mrow><mml:mi mathvariant="bold">Ψ</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, <inline-formula><mml:math id="m76"><mml:msub><mml:mrow><mml:mi mathvariant="bold">Λ</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> and <inline-formula><mml:math id="m77"><mml:msub><mml:mrow><mml:mi mathvariant="bold">Θ</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="bold-italic">g</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> (see Equation (6) in <xref ref-type="bibr" rid="r29">Perez Alonso et al., 2024</xref>).</p><fig id="f2" position="anchor" fig-type="figure" orientation="portrait"><label>Figure 2</label><caption>
<title>The Data-Generating Model With Exogenous Factors F1 and F2 and Endogenous Factors F3 and F4</title></caption><graphic xlink:href="meth.16411-f2" position="anchor" orientation="portrait"/></fig>
<p>The size of the regression parameters was set to <inline-formula><mml:math id="m78"><mml:mi>β</mml:mi></mml:math></inline-formula> and, as shown in <xref ref-type="fig" rid="f3">Figure 3</xref>, the differences between clusters were introduced by setting one regression parameter to zero in each cluster. Hence, larger values of <inline-formula><mml:math id="m79"><mml:mi>β</mml:mi></mml:math></inline-formula> resulted in larger differences and thus in greater separation between clusters, which should make the clusters easier to recover.</p><fig id="f3" position="anchor" fig-type="figure" orientation="portrait"><label>Figure 3</label><caption>
<title>The Cluster-Specific Structural Relations</title></caption><graphic xlink:href="meth.16411-f3" position="anchor" orientation="portrait"/></fig>
<p>For the group-and-cluster-specific residual factor covariances <inline-formula><mml:math id="m80"><mml:msub><mml:mrow><mml:mi mathvariant="bold">Ψ</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, we sampled the variances of the exogenous factors <inline-formula><mml:math id="m81"><mml:mi>F</mml:mi><mml:mn>1</mml:mn><mml:mi> </mml:mi></mml:math></inline-formula>and <inline-formula><mml:math id="m82"><mml:mi>F</mml:mi><mml:mn>2</mml:mn><mml:mi> </mml:mi></mml:math></inline-formula>from a uniform distribution <inline-formula><mml:math id="m83"><mml:mi>U</mml:mi><mml:mo>(</mml:mo><mml:mn>0.75</mml:mn><mml:mo>,</mml:mo><mml:mi> </mml:mi><mml:mn>1.25</mml:mn><mml:mo>)</mml:mo></mml:math></inline-formula> and their covariance from <inline-formula><mml:math id="m84"><mml:mi>U</mml:mi><mml:mo>(</mml:mo><mml:mo>-</mml:mo><mml:mn>0.3</mml:mn><mml:mo>,</mml:mo><mml:mi> </mml:mi><mml:mn>0.3</mml:mn><mml:mo>)</mml:mo></mml:math></inline-formula>. The total variances of the endogenous factors <inline-formula><mml:math id="m85"><mml:mi>F</mml:mi><mml:mn>3</mml:mn><mml:mi> </mml:mi></mml:math></inline-formula>and <inline-formula><mml:math id="m86"><mml:mi>F</mml:mi><mml:mn>4</mml:mn><mml:mi> </mml:mi></mml:math></inline-formula>were also sampled from <inline-formula><mml:math id="m87"><mml:mi>U</mml:mi><mml:mo>(</mml:mo><mml:mn>0.75</mml:mn><mml:mo>,</mml:mo><mml:mi> </mml:mi><mml:mn>1.25</mml:mn><mml:mo>)</mml:mo></mml:math></inline-formula> and their residual variances are determined as follows: For <inline-formula><mml:math id="m88"><mml:mi>F</mml:mi><mml:mn>3</mml:mn></mml:math></inline-formula> and <italic>F</italic>4, it was computed as <inline-formula><mml:math id="m89"><mml:mi mathvariant="normal">V</mml:mi><mml:mi mathvariant="normal">a</mml:mi><mml:mi mathvariant="normal">r</mml:mi><mml:msub><mml:mrow><mml:mfenced separators="|"><mml:mrow><mml:mi>F</mml:mi><mml:mn>3</mml:mn></mml:mrow></mml:mfenced></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub><mml:mo>-</mml:mo><mml:mo>(</mml:mo><mml:msubsup><mml:mrow><mml:mi>β</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn><mml:mo>,</mml:mo><mml:mi>k</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msubsup><mml:mi mathvariant="normal">V</mml:mi><mml:mi mathvariant="normal">a</mml:mi><mml:mi mathvariant="normal">r</mml:mi><mml:msub><mml:mrow><mml:mfenced separators="|"><mml:mrow><mml:mi>F</mml:mi><mml:mn>1</mml:mn></mml:mrow></mml:mfenced></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub><mml:mo>+</mml:mo><mml:msubsup><mml:mrow><mml:mi>β</mml:mi></mml:mrow><mml:mrow><mml:mn>3</mml:mn><mml:mo>,</mml:mo><mml:mi>k</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msubsup><mml:mi mathvariant="normal">V</mml:mi><mml:mi mathvariant="normal">a</mml:mi><mml:mi mathvariant="normal">r</mml:mi><mml:msub><mml:mrow><mml:mfenced separators="|"><mml:mrow><mml:mi>F</mml:mi><mml:mn>2</mml:mn></mml:mrow></mml:mfenced></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub><mml:mo>+</mml:mo><mml:mn>2</mml:mn><mml:msub><mml:mrow><mml:mi>β</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn><mml:mo>,</mml:mo><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:msub><mml:mrow><mml:mi>β</mml:mi></mml:mrow><mml:mrow><mml:mn>3</mml:mn><mml:mo>,</mml:mo><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:msub><mml:mrow><mml:mi mathvariant="normal">C</mml:mi><mml:mi mathvariant="normal">o</mml:mi><mml:mi mathvariant="normal">v</mml:mi><mml:mo>(</mml:mo><mml:mi>F</mml:mi><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mi>F</mml:mi><mml:mn>2</mml:mn><mml:mo>)</mml:mo></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> and <inline-formula><mml:math id="m90"><mml:mi mathvariant="normal">V</mml:mi><mml:mi mathvariant="normal">a</mml:mi><mml:mi mathvariant="normal">r</mml:mi><mml:msub><mml:mrow><mml:mfenced separators="|"><mml:mrow><mml:mi>F</mml:mi><mml:mn>4</mml:mn></mml:mrow></mml:mfenced></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub><mml:mo>-</mml:mo><mml:mo>(</mml:mo><mml:msubsup><mml:mrow><mml:mi>β</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mi>k</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msubsup><mml:mi mathvariant="normal">V</mml:mi><mml:mi mathvariant="normal">a</mml:mi><mml:mi mathvariant="normal">r</mml:mi><mml:msub><mml:mrow><mml:mfenced separators="|"><mml:mrow><mml:mi>F</mml:mi><mml:mn>1</mml:mn></mml:mrow></mml:mfenced></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub><mml:mo>+</mml:mo><mml:msubsup><mml:mrow><mml:mi>β</mml:mi></mml:mrow><mml:mrow><mml:mn>4</mml:mn><mml:mo>,</mml:mo><mml:mi>k</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msubsup><mml:mi mathvariant="normal">V</mml:mi><mml:mi mathvariant="normal">a</mml:mi><mml:mi mathvariant="normal">r</mml:mi><mml:msub><mml:mrow><mml:mfenced separators="|"><mml:mrow><mml:mi>F</mml:mi><mml:mn>3</mml:mn></mml:mrow></mml:mfenced></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub><mml:mo>+</mml:mo><mml:mn>2</mml:mn><mml:msub><mml:mrow><mml:mi>β</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:msub><mml:mrow><mml:mi>β</mml:mi></mml:mrow><mml:mrow><mml:mn>4</mml:mn><mml:mo>,</mml:mo><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:mfenced separators="|"><mml:mrow><mml:msub><mml:mrow><mml:mi>β</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn><mml:mo>,</mml:mo><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:mi mathvariant="normal">V</mml:mi><mml:mi mathvariant="normal">a</mml:mi><mml:mi mathvariant="normal">r</mml:mi><mml:msub><mml:mrow><mml:mfenced separators="|"><mml:mrow><mml:mi>F</mml:mi><mml:mn>1</mml:mn></mml:mrow></mml:mfenced></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub><mml:mo>+</mml:mo><mml:msub><mml:mrow><mml:mi>β</mml:mi></mml:mrow><mml:mrow><mml:mn>3</mml:mn><mml:mo>,</mml:mo><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:mi mathvariant="normal">C</mml:mi><mml:mi mathvariant="normal">o</mml:mi><mml:mi mathvariant="normal">v</mml:mi><mml:msub><mml:mrow><mml:mfenced separators="|"><mml:mrow><mml:mi>F</mml:mi><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mi>F</mml:mi><mml:mn>2</mml:mn></mml:mrow></mml:mfenced></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mfenced><mml:mo>)</mml:mo></mml:math></inline-formula>, respectively.</p><?figure f3?>
<p>In Loading Matrix <inline-formula><mml:math id="m91"><mml:msub><mml:mrow><mml:mi mathvariant="bold">Λ</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, the first loading of each factor was fixed to one. The other loadings (except for crossloadings) were approximately invariant across groups and were sampled from a normal distribution with a mean of <inline-formula><mml:math id="m92"><mml:msqrt><mml:mn>0.6</mml:mn></mml:msqrt></mml:math></inline-formula> in case of a high item reliability and <inline-formula><mml:math id="m93"><mml:msqrt><mml:mn>0.4</mml:mn></mml:msqrt></mml:math></inline-formula> in case of a low item reliability, and a variance that depended on the level of the AMI. According to a summary by <xref ref-type="bibr" rid="r3">Arts et al. (2021)</xref>, prior variances used in simulation and empirical studies typically range from 0.001 to 0.1. Following this, our simulations included five levels of AMI (i.e., 0.001, 0.005, 0.01, 0.05, 0.1). For instance, to obtain an AMI level of 0.01, which implies a variance of 0.01 for differences in loadings, we sampled loadings from a normal distribution with a variance of 0.005 for all groups.<xref ref-type="fn" rid="fn2"><sup>2</sup></xref><fn id="fn2"><label>2</label>
<p>In <italic>blavaan</italic>, the estimate of a parameter in the first group is used as the mean of the prior for that same parameter in the other groups. Consequently, the prior reflects the differences of the other groups to the reference group. The variance of the difference between two factor loadings equals the sum of their individual variances, assuming there is no covariance between them. For all groups, including the reference group, we sampled loadings from a normal distribution with a variance that is half the targeted MI level for all groups, so that the variance of the loading differences toward the reference group equals the targeted MI level.</p></fn> Per factor, one crossloading was added to the third item measuring the next factor (i.e., Item 8 crossloaded on Factor 1, Item 13 on Factor 2, Item 18 on Factor 3, and Item 3 on Factor 4). A value of 0 corresponded to no crossloading, 0.2 to a moderate crossloading, and 0.4 to a large crossloading. The unique variances on the diagonal of <inline-formula><mml:math id="m94"><mml:msub><mml:mrow><mml:mi mathvariant="bold">Θ</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="bold-italic">g</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> were sampled from <inline-formula><mml:math id="m95"><mml:mi>U</mml:mi><mml:mo>(</mml:mo><mml:mn>0.3</mml:mn><mml:mo>,</mml:mo><mml:mi> </mml:mi><mml:mn>0.5</mml:mn><mml:mo>)</mml:mo></mml:math></inline-formula> under high reliability and from <inline-formula><mml:math id="m96"><mml:mi>U</mml:mi><mml:mo>(</mml:mo><mml:mn>0.50</mml:mn><mml:mo>,</mml:mo><mml:mi> </mml:mi><mml:mn>0.70</mml:mn><mml:mo>)</mml:mo></mml:math></inline-formula> under low reliability.</p>
<p>Finally, the data were sampled from <inline-formula><mml:math id="m97"><mml:mi mathvariant="normal">M</mml:mi><mml:mi mathvariant="normal">V</mml:mi><mml:mi mathvariant="normal">N</mml:mi><mml:mo>(</mml:mo><mml:mn>0</mml:mn><mml:mo>,</mml:mo><mml:mi mathvariant="bold-italic"> </mml:mi><mml:msub><mml:mrow><mml:mi mathvariant="bold">Σ</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:mo>)</mml:mo></mml:math></inline-formula> for each group. In total, we generated 2 (number of groups) × 3 (within-group sample size) × 2 (number of clusters) × 3 (size of regression parameters) × 2 (reliability) × 5 (size of AMI) × 3 (size of crossloadings) × 50 (replications) = 54,000 data sets according to the described procedure, using R Version 4.4 (<xref ref-type="bibr" rid="r32">R Core Team, 2022</xref>). All data sets were analyzed with MixMG-BSEM with 50 random starts and the true number of clusters. For each data set, we performed the analysis five times, with different prior variances for the loadings (i.e., 0.001, 0.005, 0.01, 0.05, 0.1) in Step 1, to examine the performance of MixMG-BSEM across different prior variances. The analyses were performed on a supercomputer consisting of 2 Intel Xeon Platinum 8468 CPUs (Sapphire Rapids). The average computation time was 50.5 minutes (<inline-formula><mml:math id="m98"><mml:mi>S</mml:mi><mml:mi>D</mml:mi><mml:mo>=</mml:mo><mml:mi> </mml:mi><mml:mn>24.8</mml:mn></mml:math></inline-formula>) for Step 1 (mainly influenced by <inline-formula><mml:math id="m99"><mml:mi>G</mml:mi></mml:math></inline-formula> and <inline-formula><mml:math id="m100"><mml:msub><mml:mrow><mml:mi>N</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>) with the four factors estimated sequentially, 0.03 minutes (<inline-formula><mml:math id="m101"><mml:mi>S</mml:mi><mml:mi>D</mml:mi><mml:mo>=</mml:mo><mml:mn>0.02</mml:mn></mml:math></inline-formula>) for the intermediate Step 2, and 5.7 minutes (<inline-formula><mml:math id="m102"><mml:mi>S</mml:mi><mml:mi>D</mml:mi><mml:mo>=</mml:mo><mml:mi> </mml:mi><mml:mn>5.7</mml:mn></mml:math></inline-formula>) for Step 3 (mainly influenced by <inline-formula><mml:math id="m103"><mml:msub><mml:mrow><mml:mi>N</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, <inline-formula><mml:math id="m104"><mml:mi>K</mml:mi></mml:math></inline-formula>, and <inline-formula><mml:math id="m105"><mml:mi>β</mml:mi><mml:mo>)</mml:mo></mml:math></inline-formula>.<xref ref-type="fn" rid="fn3"><sup>3</sup></xref><fn id="fn3"><label>3</label>
		<p>The first step of MixMG-BSEM (i.e., estimating the MM using <italic>blavaan</italic>) can be computationally demanding, especially for larger sample sizes. Luckily, the stepwise estimation of MixMG-BSEM implies that the MM needs to be estimated only once, even when estimating the SM with different numbers of clusters for model selection. To illustrate, we report computation times for one of the largest data sets in our simulation study, which included 48 groups and 200 observations per group. Using <italic>blavaan</italic>, estimating the four factors sequentially without parallel computing for the MCMC chains (per factor) took around 62 minutes in total. With parallel computing applied to the three MCMC chains (per factor), the computation time decreased to 42 minutes. Note that additional speed gains could be achieved by further parallelizing across the four factors, depending on hardware capabilities and user preferences. Alternatively, Mplus offers a more time-efficient estimation of the MM with AMI, though it is commercial software. For the same data set, <italic>Mplus</italic> completed the estimation of the four factors sequentially in only 2 minutes. It is worth noting that eliminating the mean structure by centering per group (see <xref ref-type="sec" rid="M1">Method</xref> section) helped as well, since these computation times of <italic>blavaan</italic> and <italic>Mplus</italic> increased to 66 and 91 minutes, respectively, when including the mean structure in the model.</p></fn> We assessed MCMC convergence of the Bayesian estimation using the potential scale reduction factor (PSRF). On average, the PSRF was 1.000 (<inline-formula><mml:math id="m106"><mml:mi>S</mml:mi><mml:mi>D</mml:mi><mml:mi> </mml:mi><mml:mo>=</mml:mo><mml:mi> </mml:mi><mml:mn>0.001</mml:mn></mml:math></inline-formula>), indicating sufficient convergence, with no notable differences across conditions. These results suggest that the MCMC chains converged properly for all simulated data sets.</p></sec>
<sec sec-type="results"><title>Results</title>
<sec><title>Recovery of Factor Loadings</title>
	<p>We evaluated the recovery of the group-specific factor loading estimates for each item <inline-formula><mml:math id="m107"><mml:mi>j</mml:mi></mml:math></inline-formula>, using the Root Mean Squared Error (RMSE) across groups as follows:</p><disp-formula id="eqn-6"><label>3</label><mml:math id="m108" display="block"><mml:mtable><mml:mtr><mml:mtd><mml:mrow><mml:maligngroup/><mml:msub><mml:mrow><mml:mi>R</mml:mi><mml:mi>M</mml:mi><mml:mi>S</mml:mi><mml:mi>E</mml:mi></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>λ</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:msqrt><mml:mfrac><mml:mrow><mml:mrow><mml:msubsup><mml:mo stretchy="false">∑</mml:mo><mml:mrow><mml:mi>g</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>G</mml:mi></mml:mrow></mml:msubsup><mml:mrow><mml:msup><mml:mrow><mml:mfenced separators="|"><mml:mrow><mml:msub><mml:mrow><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>λ</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>g</mml:mi><mml:mi>j</mml:mi></mml:mrow></mml:msub><mml:mo>-</mml:mo><mml:mi>λ</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mfenced></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup></mml:mrow></mml:mrow></mml:mrow><mml:mrow><mml:mi>G</mml:mi></mml:mrow></mml:mfrac></mml:msqrt><mml:mi> </mml:mi></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>where <inline-formula><mml:math id="m109"><mml:msub><mml:mrow><mml:mi>λ</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> is the true group-specific loading of the <italic>j-</italic>th item on the factor, and <inline-formula><mml:math id="m110"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>λ</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>g</mml:mi><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> is the corresponding estimate. Note that <inline-formula><mml:math id="m111"><mml:msub><mml:mrow><mml:mi>R</mml:mi><mml:mi>M</mml:mi><mml:mi>S</mml:mi><mml:mi>E</mml:mi></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>λ</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:msub></mml:math></inline-formula> is also affected by the variance of the estimates, rather than just the bias.</p>
<p>When using MixMG-BSEM with the true prior variances for the loadings, the average <inline-formula><mml:math id="m112"><mml:msub><mml:mrow><mml:mi>R</mml:mi><mml:mi>M</mml:mi><mml:mi>S</mml:mi><mml:mi>E</mml:mi></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>λ</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:msub></mml:math></inline-formula> across the four factors and all simulated data sets was 0.066, 0.091, 0.066, and 0.066, respectively, for the loadings of the second to the fifth item of each factor (<xref ref-type="table" rid="t1">Table 1</xref>, last row). Note that <inline-formula><mml:math id="m113"><mml:mi>R</mml:mi><mml:mi>M</mml:mi><mml:msub><mml:mrow><mml:mi>S</mml:mi><mml:mi>E</mml:mi></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>λ</mml:mi></mml:mrow><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:msub></mml:math></inline-formula> was larger due to the disregarded crossloadings on that item. This was also the only <inline-formula><mml:math id="m114"><mml:mi>R</mml:mi><mml:mi>M</mml:mi><mml:mi>S</mml:mi><mml:msub><mml:mrow><mml:mi>E</mml:mi></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>λ</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:msub></mml:math></inline-formula> value that differed across the four factors. Specifically, the <inline-formula><mml:math id="m115"><mml:mi>R</mml:mi><mml:mi>M</mml:mi><mml:mi>S</mml:mi><mml:msub><mml:mrow><mml:mi>E</mml:mi></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>λ</mml:mi></mml:mrow><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:msub></mml:math></inline-formula> values were 0.093, 0.072, 0.098, and 0.101, for <inline-formula><mml:math id="m116"><mml:mi>F</mml:mi><mml:mn>1</mml:mn></mml:math></inline-formula> to <inline-formula><mml:math id="m117"><mml:mi>F</mml:mi><mml:mn>4</mml:mn></mml:math></inline-formula>, respectively. It seems that the third loading for <inline-formula><mml:math id="m118"><mml:mi>F</mml:mi><mml:mn>2</mml:mn></mml:math></inline-formula> is affected by the ignored crossloading the least, which may be explained by the fact that, unlike the other factors, <inline-formula><mml:math id="m119"><mml:mi>F</mml:mi><mml:mn>2</mml:mn></mml:math></inline-formula> is involved in only one direct regression relation with the other factors<xref ref-type="fn" rid="fn4"><sup>4</sup></xref><fn id="fn4"><label>4</label>
<p>It has indirect relations with the other factors via the correlation between F1 and F2, but the expected value of this correlation is zero.</p></fn> and is thus less correlated with the other factors. When the crossloadings were zero (i.e., without crossloadings), <inline-formula><mml:math id="m120"><mml:msub><mml:mrow><mml:mi>R</mml:mi><mml:mi>M</mml:mi><mml:mi>S</mml:mi><mml:mi>E</mml:mi></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>λ</mml:mi></mml:mrow><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:msub></mml:math></inline-formula> took on the same values as for the other loadings (<inline-formula><mml:math id="m121"><mml:msub><mml:mrow><mml:mi>R</mml:mi><mml:mi>M</mml:mi><mml:mi>S</mml:mi><mml:mi>E</mml:mi></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>λ</mml:mi></mml:mrow><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mn>0.066</mml:mn></mml:math></inline-formula>), whereas they increased with larger crossloadings (see <xref ref-type="table" rid="t1">Table 1</xref>). <inline-formula><mml:math id="m122"><mml:mi>R</mml:mi><mml:mi>M</mml:mi><mml:mi>S</mml:mi><mml:msub><mml:mrow><mml:mi>E</mml:mi></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>λ</mml:mi></mml:mrow><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:msub></mml:math></inline-formula> was also higher in case of larger <inline-formula><mml:math id="m123"><mml:mi>β</mml:mi></mml:math></inline-formula>, which implies stronger correlations between factors (see <xref ref-type="table" rid="t1">Table 1</xref>). Note that larger <inline-formula><mml:math id="m124"><mml:msub><mml:mrow><mml:mi>N</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo></mml:math></inline-formula> higher item reliability, and lower levels of AMI — thus applying lower prior variances — resulted in lower <inline-formula><mml:math id="m125"><mml:mi>R</mml:mi><mml:mi>M</mml:mi><mml:mi>S</mml:mi><mml:msub><mml:mrow><mml:mi>E</mml:mi></mml:mrow><mml:mrow><mml:mi>λ</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> for all items. The latter is explained by the fact that a lower prior variance more strongly approximates an equality constraint, which lowers the sample size requirements and thus the estimates’ variability for a given sample size.</p>
<table-wrap id="t1" position="anchor" orientation="portrait">
<label>Table 1</label><caption><title>The Average<inline-formula><mml:math id="m126"><mml:mi mathvariant="normal"> </mml:mi><mml:mi mathvariant="normal">R</mml:mi><mml:mi mathvariant="normal">M</mml:mi><mml:mi mathvariant="normal">S</mml:mi><mml:msub><mml:mrow><mml:mi>E</mml:mi></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>λ</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:msub></mml:math></inline-formula> for Factor Loading Estimates When Using the True Prior Variances for the Loadings</title></caption>
<table frame="hsides" rules="groups">
<col width="" align="left"/>
<col width=""/>
<col width=""/>
<col width=""/>
<col width=""/>
<col width=""/>
<thead>
<tr>
<th>Factor</th>
<th>Level</th>
	<th><inline-formula><mml:math id="m127"><mml:msub><mml:mrow><mml:mi>R</mml:mi><mml:mi>M</mml:mi><mml:mi>S</mml:mi><mml:mi>E</mml:mi></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>λ</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:msub></mml:math></inline-formula> (<italic>SD</italic>)</th>
	<th><inline-formula><mml:math id="m128"><mml:msub><mml:mrow><mml:mi>R</mml:mi><mml:mi>M</mml:mi><mml:mi>S</mml:mi><mml:mi>E</mml:mi></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>λ</mml:mi></mml:mrow><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:msub></mml:math></inline-formula> (<italic>SD</italic>)</th>
	<th><inline-formula><mml:math id="m129"><mml:msub><mml:mrow><mml:mi>R</mml:mi><mml:mi>M</mml:mi><mml:mi>S</mml:mi><mml:mi>E</mml:mi></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>λ</mml:mi></mml:mrow><mml:mrow><mml:mn>4</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:msub></mml:math></inline-formula> (<italic>SD</italic>)</th>
	<th><inline-formula><mml:math id="m130"><mml:msub><mml:mrow><mml:mi>R</mml:mi><mml:mi>M</mml:mi><mml:mi>S</mml:mi><mml:mi>E</mml:mi></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>λ</mml:mi></mml:mrow><mml:mrow><mml:mn>5</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:msub></mml:math></inline-formula> (<italic>SD</italic>)</th>
</tr>
</thead>
<tbody>
<tr style="grey-border-top">
<td><inline-formula><mml:math id="m131"><mml:mi>G</mml:mi></mml:math></inline-formula></td>
<td>24</td>
<td>0.066 (0.031)</td>
<td>0.091 (0.041)</td>
<td>0.066 (0.031)</td>
<td>0.066 (0.031)</td>
</tr>
<tr>
<td/>	
<td>48</td>
<td>0.065 (0.031)</td>
<td>0.090 (0.041)</td>
<td>0.065 (0.031)</td>
<td>0.065 (0.031)</td>
</tr>
<tr style="grey-border-top">
<td><inline-formula><mml:math id="m132"><mml:msub><mml:mrow><mml:mi>N</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula></td>
<td>50</td>
<td>0.079 (0.038)</td>
<td>0.102 (0.046)</td>
<td>0.079 (0.038)</td>
<td>0.079 (0.038)</td>
</tr>
<tr>
<td/>	
<td>100</td>
<td>0.065 (0.028)</td>
<td>0.091 (0.039)</td>
<td>0.065 (0.028)</td>
<td>0.065 (0.028)</td>
</tr>
<tr>
<td/>	
<td>200</td>
<td>0.052 (0.019)</td>
<td>0.081 (0.035)</td>
<td>0.052 (0.019)</td>
<td>0.052 (0.019)</td>
</tr>
<tr style="grey-border-top">
<td><inline-formula><mml:math id="m133"><mml:mi>K</mml:mi></mml:math></inline-formula></td>
<td>2</td>
<td>0.066 (0.031)</td>
<td>0.091 (0.041)</td>
<td>0.066 (0.031)</td>
<td>0.066 (0.031)</td>
</tr>
<tr>
<td/>	
<td>4</td>
<td>0.066 (0.031)</td>
<td>0.091 (0.041)</td>
<td>0.066 (0.031)</td>
<td>0.066 (0.031)</td>
</tr>
<tr style="grey-border-top"><?pagebreak-before?>
<td><inline-formula><mml:math id="m134"><mml:mi>β</mml:mi></mml:math></inline-formula></td>
<td align="char" char=".">0.2</td>
<td>0.066 (0.031)</td>
<td>0.080 (0.035)</td>
<td>0.066 (0.031)</td>
<td>0.066 (0.031)</td>
</tr>
<tr>
<td/>	
<td align="char" char=".">0.3</td>
<td>0.066 (0.031)</td>
<td>0.090 (0.039)</td>
<td>0.066 (0.031)</td>
<td>0.066 (0.031)</td>
</tr>
<tr>
<td/>	
<td align="char" char=".">0.4</td>
<td>0.065 (0.031)</td>
<td>0.103 (0.045)</td>
<td>0.065 (0.031)</td>
<td>0.065 (0.031)</td>
</tr>
<tr style="grey-border-top">
<td>Reliability</td>
<td>high</td>
<td>0.061 (0.028)</td>
<td>0.087 (0.039)</td>
<td>0.061 (0.027)</td>
<td>0.061 (0.027)</td>
</tr>
<tr>
<td/>
<td>low</td>
<td>0.070 (0.034)</td>
<td>0.095 (0.043)</td>
<td>0.070 (0.034)</td>
<td>0.070 (0.034)</td>
</tr>
<tr style="grey-border-top">
<td>AMI</td>
<td align="char" char=".">0.001</td>
<td>0.028 (0.006)</td>
<td>0.055 (0.029)</td>
<td>0.028 (0.006)</td>
<td>0.028 (0.006)</td>
</tr>
<tr>
<td/>	
<td align="char" char=".">0.005</td>
<td>0.048 (0.006)</td>
<td>0.072 (0.027)</td>
<td>0.048 (0.006)</td>
<td>0.048 (0.006)</td>
</tr>
<tr>
<td/>
<td align="char" char=".">0.01</td>
<td>0.060 (0.009)</td>
<td>0.084 (0.027)</td>
<td>0.060 (0.009)</td>
<td>0.060 (0.009)</td>
</tr>
<tr>
<td/>
<td align="char" char=".">0.05</td>
<td>0.090 (0.020)</td>
<td>0.116 (0.032)</td>
<td>0.090 (0.020)</td>
<td>0.090 (0.020)</td>
</tr>
<tr>
<td/>
<td align="char" char=".">0.1</td>
<td>0.102 (0.027)</td>
<td>0.128 (0.037)</td>
<td>0.102 (0.026)</td>
<td>0.102 (0.027)</td>
</tr>
<tr style="grey-border-top">
<td>Crossloadings</td>
<td>0</td>
<td>0.066 (0.031)</td>
<td>0.066 (0.031)</td>
<td>0.066 (0.031)</td>
<td>0.066 (0.031)</td>
</tr>
<tr>
<td/>
<td align="char" char=".">0.2</td>
<td>0.066 (0.031)</td>
<td>0.086 (0.032)</td>
<td>0.066 (0.031)</td>
<td>0.065 (0.031)</td>
</tr>
<tr>
<td/>	
<td align="char" char=".">0.4</td>
<td>0.066 (0.031)</td>
<td>0.122 (0.037)</td>
<td>0.066 (0.031)</td>
<td>0.066 (0.031)</td>
</tr>
	<tr style="background-lightblue; white-border-top; white-border-bottom">
<td align="left" colspan="2">Total</td>
<td>0.066 (0.031)</td>
<td>0.091 (0.041)</td>
<td>0.066 (0.031)</td>
<td>0.066 (0.031)</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>To illustrate the effect of the prior variances for the loadings, <inline-formula><mml:math id="m135"><mml:msub><mml:mrow><mml:mi>R</mml:mi><mml:mi>M</mml:mi><mml:mi>S</mml:mi><mml:mi>E</mml:mi></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>λ</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:msub></mml:math></inline-formula> across different prior variances is shown in <xref ref-type="fig" rid="f4">Figure 4</xref>. The diagonal of the plot represents cases where the prior variances were correctly specified, while the lower part shows cases where the priors were narrower than the true level of AMI. In general, applying too narrow or too wide priors resulted in larger <inline-formula><mml:math id="m136"><mml:mi>R</mml:mi><mml:mi>M</mml:mi><mml:mi>S</mml:mi><mml:msub><mml:mrow><mml:mi>E</mml:mi></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>λ</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:msub></mml:math></inline-formula> values. Since the prior variance affected the loading recovery, we also evaluated prior selection using the Deviance Information Criterion (DIC), which balances model fit and complexity. When looking at the prior selection per loading, the correct selection rate was 59.2% across all loadings and all simulated data sets.<xref ref-type="fn" rid="fn5"><sup>5</sup></xref><fn id="fn5"><label>5</label>
<p>Similar results were found with the widely applicable information criterion (WAIC; <xref ref-type="bibr" rid="r39">Watanabe, 2010</xref>) and leave-one-out information criterion (LOOIC; <xref ref-type="bibr" rid="r37">Vehtari et al., 2017</xref>): WAIC: 58.2%; LOOIC: 57.6%.</p></fn> For 28.2% of the data sets, the prior selection was flawless in the sense that true priors were selected for <italic>all</italic> loadings. Generally, the DIC tended to select either the true or slightly smaller prior variances. Specifically, for an AMI level of 0.001, DIC correctly selected the prior variance of 0.001 for 86.6% of the loadings, with smaller proportions selecting 0.005 (12.5%) and 0.01 (0.9%). For an AMI level of 0.005, the correct selection rate was 59.4%, followed by 0.001 (36.6%) and 0.01 (4.0%). For an AMI level of 0.01, DIC most frequently selected 0.005 (71.1%), followed by 0.01 (23.5%) and 0.001 (5.4%). For an AMI level of 0.05, DIC primarily selected the true prior variance (81.5%), followed by prior variances of 0.1 (10.0%) and 0.01 (8.5%). For an AMI level of 0.1, DIC mostly selected prior variances of 0.05 (55.1%) and 0.1 (44.9%). Although prior selection based on the DIC varied across different levels of AMI and was not always optimal, the resulting <inline-formula><mml:math id="m137"><mml:mi>R</mml:mi><mml:mi>M</mml:mi><mml:mi>S</mml:mi><mml:msub><mml:mrow><mml:mi>E</mml:mi></mml:mrow><mml:mrow><mml:mi>λ</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> values remained relatively stable given the selected priors, suggesting stable factor loading recovery despite prior variance misspecification.</p><fig id="f4" position="anchor" fig-type="figure" orientation="portrait"><label>Figure 4</label><caption>
<title><inline-formula><mml:math id="m138"><mml:msub><mml:mrow><mml:mi>R</mml:mi><mml:mi>M</mml:mi><mml:mi>S</mml:mi><mml:mi>E</mml:mi></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>λ</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:msub></mml:math></inline-formula> Across Different Prior Variances, Indicated by the Columns, Whereas the Rows Represent the True Levels of AMI</title><p><italic>Note.</italic> The diagonal (in white) contains cases where the prior variances were correctly specified, while the lower part represents cases where the priors were too narrow. For each row, the cells are colored red if the <inline-formula><mml:math id="m139"><mml:msub><mml:mrow><mml:mi>R</mml:mi><mml:mi>M</mml:mi><mml:mi>S</mml:mi><mml:mi>E</mml:mi></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>λ</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:msub></mml:math></inline-formula> is larger than the <inline-formula><mml:math id="m140"><mml:msub><mml:mrow><mml:mi>R</mml:mi><mml:mi>M</mml:mi><mml:mi>S</mml:mi><mml:mi>E</mml:mi></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>λ</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:msub></mml:math></inline-formula> on the diagonal, and blue if it is smaller.</p></caption><graphic xlink:href="meth.16411-f4" position="anchor" orientation="portrait"/></fig></sec>
<sec><title>Sensitivity to Local Maxima</title>
<p>To evaluate how often (Step 3 of) MixMG-BSEM converged to a local maximum, we compared the log-likelihood of the final best solution (out of 50 random starts) to the one obtained when starting from the true clustering, which is a proxy for the global maximum. If <inline-formula><mml:math id="m141"><mml:mrow><mml:mrow><mml:mi mathvariant="normal">log</mml:mi></mml:mrow><mml:mo>⁡</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>L</mml:mi></mml:mrow><mml:mrow><mml:mi>η</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mrow></mml:math></inline-formula>was more than 0.001 lower than the proxy, the solution was considered a local maximum. Overall, when applying the true priors, MixMG-BSEM ended up in a local maximum for 0.01% of the data sets, with all local maxima occurring in case of <inline-formula><mml:math id="m142"><mml:msub><mml:mrow><mml:mi>N</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mn>50</mml:mn></mml:math></inline-formula> with <inline-formula><mml:math id="m143"><mml:mi>K</mml:mi><mml:mo>=</mml:mo><mml:mn>4</mml:mn></mml:math></inline-formula>.</p></sec>
<sec><title>Recovery of Clusters</title>
<p>The Adjusted Rand Index (ARI; <xref ref-type="bibr" rid="r14">Hubert &amp; Arabie, 1985</xref>) measures the similarity between two partitions while correcting for chance, with a value of one indicating perfect agreement and zero indicating the level of agreement between two random partitions. To compute the ARI, the modal clustering (i.e., assigning each group to the cluster with the highest classification probability) was compared to the true clustering. Additionally, the correct clustering rate (%CC) was computed as the percentage of correctly clustered groups.</p>
<p>When using the true priors, the average ARI across all simulated data was 0.644 and the %CC was 84.8%.<xref ref-type="fn" rid="fn6"><sup>6</sup></xref><fn id="fn6"><label>6</label>
		<p>We further examined the distributions of the ARI using boxplots across relevant conditions (see Figure 1 in the Supplementary Material, S3; <xref ref-type="bibr" rid="r44">Zhao et al., 2025b</xref>). Under more challenging conditions (e.g., small sample sizes, low reliability, small regression effects), we observed larger variability. For example, with <inline-formula><mml:math id="m144"><mml:mi>β</mml:mi><mml:mo>=</mml:mo><mml:mn>0.2</mml:mn></mml:math></inline-formula> and <inline-formula><mml:math id="m145"><mml:msub><mml:mrow><mml:mi>N</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mn>50</mml:mn></mml:math></inline-formula>, the distributions were more right-skewed, indicating a majority of low ARI values. In contrast, under easier conditions, such as <inline-formula><mml:math id="m146"><mml:mi>β</mml:mi><mml:mo>=</mml:mo><mml:mn>0.4</mml:mn></mml:math></inline-formula> and <inline-formula><mml:math id="m147"><mml:msub><mml:mrow><mml:mi>N</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub><mml:mi> </mml:mi><mml:mo>=</mml:mo><mml:mn>200</mml:mn></mml:math></inline-formula>, the distributions became symmetric or slightly left-skewed, with less variability, indicating more consistent high ARI values. Therefore, we have added a table reporting the medians and MADs for ARI and %CC in the Supplementary Material (S4; <xref ref-type="bibr" rid="r44">Zhao et al., 2025b</xref>), to complement the means and <italic>SD</italic>s reported in the main text.</p></fn> To check which main effects and two-way interactions among the manipulated factors significantly influenced the ARI, we conducted an analysis of variance (ANOVA) using the <italic>aov</italic> function in R. The ANOVA results table is provided in Supplementary Material (S1; <xref ref-type="bibr" rid="r44">Zhao et al., 2025b</xref>). Firstly, all main effects were significant at the <inline-formula><mml:math id="m148"><mml:mi>α</mml:mi><mml:mo>=</mml:mo><mml:mn>0.01</mml:mn></mml:math></inline-formula> level. From <xref ref-type="table" rid="t2">Table 2</xref>, we see that larger <inline-formula><mml:math id="m149"><mml:mi>G</mml:mi></mml:math></inline-formula>, larger <inline-formula><mml:math id="m150"><mml:msub><mml:mrow><mml:mi>N</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, smaller <inline-formula><mml:math id="m151"><mml:mi>K</mml:mi></mml:math></inline-formula>, larger <inline-formula><mml:math id="m152"><mml:mi>β</mml:mi></mml:math></inline-formula>, higher reliability, all led to better recovery of clusters, with <inline-formula><mml:math id="m153"><mml:msub><mml:mrow><mml:mi>N</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> and <inline-formula><mml:math id="m154"><mml:mi>β</mml:mi></mml:math></inline-formula> having the strongest effect, as reflected in higher ARI and %CC. The differences in ARI across levels of AMI and crossloadings were very small, suggesting that cluster recovery was relatively unaffected by these factors. Secondly, <inline-formula><mml:math id="m155"><mml:msub><mml:mrow><mml:mi>N</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, <inline-formula><mml:math id="m156"><mml:mi>K</mml:mi></mml:math></inline-formula>, <inline-formula><mml:math id="m157"><mml:mi>β</mml:mi></mml:math></inline-formula> and reliability all interacted significantly with one another (i.e., all six of their two-way interactions were significant), which is why we further illustrate the four-way interaction effect of these four manipulated factors on ARI in <xref ref-type="fig" rid="f5">Figure 5</xref>. We see that the ARI was very sensitive to <inline-formula><mml:math id="m158"><mml:msub><mml:mrow><mml:mi>N</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, and more so in case of <inline-formula><mml:math id="m159"><mml:mi>K</mml:mi><mml:mo>=</mml:mo><mml:mn>4</mml:mn></mml:math></inline-formula>, smaller <inline-formula><mml:math id="m160"><mml:mi>β</mml:mi></mml:math></inline-formula>, or low reliability. According to <xref ref-type="bibr" rid="r36">Steinley (2004)</xref>, ARI values above 0.80 indicate good cluster recovery. Using this rule-of-thumb, the cluster recovery was good when <inline-formula><mml:math id="m161"><mml:mi>β</mml:mi><mml:mo>=</mml:mo><mml:mn>0.4</mml:mn></mml:math></inline-formula> with <inline-formula><mml:math id="m162"><mml:msub><mml:mrow><mml:mi>N</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub><mml:mo>≥</mml:mo><mml:mn>100</mml:mn></mml:math></inline-formula>, or when <inline-formula><mml:math id="m163"><mml:mi>β</mml:mi><mml:mo>=</mml:mo><mml:mn>0.3</mml:mn></mml:math></inline-formula> with <inline-formula><mml:math id="m164"><mml:msub><mml:mrow><mml:mi>N</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mn>200</mml:mn><mml:mo>.</mml:mo></mml:math></inline-formula> When <inline-formula><mml:math id="m165"><mml:mi>β</mml:mi><mml:mo>=</mml:mo><mml:mn>0.2</mml:mn></mml:math></inline-formula>, the ARI only exceeded 0.80 with <inline-formula><mml:math id="m166"><mml:msub><mml:mrow><mml:mi>N</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> = 200 for two clusters and a high reliability.</p>
<table-wrap id="t2" position="anchor" orientation="portrait">
<label>Table 2</label><caption><title>The Average ARI and Correct Clustering Rate (%CC) When Using the True Prior Variances for the Loadings</title></caption>
<table frame="hsides" rules="groups">
<col width="" align="left"/>
<col width=""/>
<col width=""/>
<col width=""/>
<thead>
<tr>
<th>Factor</th>
<th>Level</th>
<th>ARI (<italic>SD</italic>)</th>
<th>%CC (<italic>SD</italic>)</th>
</tr>
</thead>
<tbody>
<tr>
<td><inline-formula><mml:math id="m167"><mml:mi>G</mml:mi></mml:math></inline-formula></td>
<td>24</td>
<td>0.628 (0.331)</td>
<td>0.840 (0.170)</td>
</tr>
<tr>
<td/>	
<td>48</td>
<td>0.660 (0.301)</td>
<td>0.856 (0.159)</td>
</tr>
<tr style="grey-border-top">
<td><inline-formula><mml:math id="m168"><mml:msub><mml:mrow><mml:mi>N</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula></td>
<td>50</td>
<td>0.402 (0.280)</td>
<td>0.733 (0.174)</td>
</tr>
<tr>
<td/>	
<td>100</td>
<td>0.663 (0.284)</td>
<td>0.861 (0.145)</td>
</tr>
<tr>
<td/>	
<td>200</td>
<td>0.865 (0.187)</td>
<td>0.949 (0.082)</td>
</tr>
<tr style="grey-border-top">
<td><inline-formula><mml:math id="m169"><mml:mi>K</mml:mi></mml:math></inline-formula></td>
<td>2</td>
<td>0.719 (0.295)</td>
<td>0.913 (0.106)</td>
</tr>
<tr>
<td/>	
<td>4</td>
<td>0.569 (0.320)</td>
<td>0.782 (0.187)</td>
</tr>
<tr style="grey-border-top">
<td><inline-formula><mml:math id="m170"><mml:mi>β</mml:mi></mml:math></inline-formula></td>
<td align="char" char=".">0.2</td>
<td>0.400 (0.282)</td>
<td>0.733 (0.175)</td>
</tr>
<tr>
<td/>	
<td align="char" char=".">0.3</td>
<td>0.685 (0.280)</td>
<td>0.871 (0.142)</td>
</tr>
<tr>
<td/>	
<td align="char" char=".">0.4</td>
<td>0.849 (0.201)</td>
<td>0.940 (0.095)</td>
</tr>
<tr style="grey-border-top">
<td>Reliability</td>
<td>high</td>
<td>0.682 (0.309)</td>
<td>0.865 (0.159)</td>
</tr>
<tr>
<td/>	
<td>low</td>
<td>0.606 (0.320)</td>
<td>0.830 (0.170)</td>
</tr>
<tr style="grey-border-top">
<td>AMI</td>
<td align="char" char=".">0.001</td>
<td>0.651 (0.314)</td>
<td>0.851 (0.163)</td>
</tr>
<tr>
<td/>	
<td align="char" char=".">0.005</td>
<td>0.647 (0.317)</td>
<td>0.849 (0.164)</td>
</tr>
<tr>
<td/>	
<td align="char" char=".">0.01</td>
<td>0.649 (0.316)</td>
<td>0.850 (0.164)</td>
</tr>
<tr>
<td/>	
<td align="char" char=".">0.05</td>
<td>0.638 (0.319)</td>
<td>0.845 (0.167)</td>
</tr>
<tr>
<td/>	
<td align="char" char=".">0.1</td>
<td>0.635 (0.318)</td>
<td>0.843 (0.166)</td>
</tr>
<tr style="grey-border-top">
<td>Crossloadings</td>
<td>0</td>
<td>0.654 (0.316)</td>
<td>0.852 (0.164)</td>
</tr>
<tr>
<td/>	
<td align="char" char=".">0.2</td>
<td>0.644 (0.318)</td>
<td>0.848 (0.166)</td>
</tr>
<tr>
<td/>	
<td align="char" char=".">0.4</td>
<td>0.634 (0.318)</td>
<td>0.844 (0.166)</td>
</tr>
<tr style="background-lightblue; white-border-top; white-border-bottom">
<td align="left" colspan="2">Total</td>
<td>0.644 (0.317)</td>
<td>0.848 (0.165)</td>
</tr>
</tbody>
</table>
</table-wrap><fig id="f5" position="anchor" fig-type="figure" orientation="portrait"><label>Figure 5</label><caption>
<title>The ARI for MixMG-BSEM in Function of the Within-Group Sample Sizes, Number of Clusters, Size of Regression Parameters, and Reliability</title></caption><graphic xlink:href="meth.16411-f5" position="anchor" orientation="portrait"/></fig>
<p>Across different prior variances, the ARI remained relatively stable. When the applied prior was too narrow or too large, the ARI slightly dropped. For example, for an AMI level of 0.1, it decreased from 0.635 when using the true prior variance to 0.608 when using a prior variance of 0.001. For an AMI level of 0.001, it decreased from 0.651 when using the true prior variance to 0.638 when using a prior variance of 0.1.</p></sec><?table t2?>
<sec><title>Recovery of Regression Parameters</title>
<p>To evaluate the recovery of the regression parameters, we computed the <inline-formula><mml:math id="m171"><mml:mi>M</mml:mi><mml:msub><mml:mrow><mml:mi>E</mml:mi></mml:mrow><mml:mrow><mml:mi>β</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> and the <inline-formula><mml:math id="m172"><mml:msub><mml:mrow><mml:mi>R</mml:mi><mml:mi>M</mml:mi><mml:mi>S</mml:mi><mml:mi>E</mml:mi></mml:mrow><mml:mrow><mml:mi>β</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> per regression parameter (i.e., <inline-formula><mml:math id="m173"><mml:msub><mml:mrow><mml:mi>β</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:math></inline-formula>, <inline-formula><mml:math id="m174"><mml:msub><mml:mrow><mml:mi>β</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:math></inline-formula>, <inline-formula><mml:math id="m175"><mml:msub><mml:mrow><mml:mi>β</mml:mi></mml:mrow><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msub></mml:math></inline-formula>, and <inline-formula><mml:math id="m176"><mml:msub><mml:mrow><mml:mi>β</mml:mi></mml:mrow><mml:mrow><mml:mn>4</mml:mn></mml:mrow></mml:msub></mml:math></inline-formula>):</p><disp-formula id="e___6"><mml:math id="m177"><mml:mtable><mml:mtr><mml:mtd><mml:mrow><mml:maligngroup/><mml:msub><mml:mrow><mml:mi>M</mml:mi><mml:mi>E</mml:mi></mml:mrow><mml:mrow><mml:mi>β</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mrow><mml:msubsup><mml:mo stretchy="false">∑</mml:mo><mml:mrow><mml:mi>k</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>K</mml:mi></mml:mrow></mml:msubsup><mml:mrow><mml:mfenced separators="|"><mml:mrow><mml:msub><mml:mrow><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>β</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:mo>-</mml:mo><mml:mi>β</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mfenced></mml:mrow></mml:mrow></mml:mrow><mml:mrow><mml:mi>K</mml:mi></mml:mrow></mml:mfrac><mml:mi> </mml:mi><mml:mo>#</mml:mo><mml:mfenced separators="|"><mml:mrow><mml:mn>7</mml:mn></mml:mrow></mml:mfenced></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula><disp-formula id="e___7"><mml:math id="m178"><mml:mtable><mml:mtr><mml:mtd><mml:mrow><mml:maligngroup/><mml:msub><mml:mrow><mml:mi>R</mml:mi><mml:mi>M</mml:mi><mml:mi>S</mml:mi><mml:mi>E</mml:mi></mml:mrow><mml:mrow><mml:mi>β</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:msqrt><mml:mfrac><mml:mrow><mml:mrow><mml:msubsup><mml:mo stretchy="false">∑</mml:mo><mml:mrow><mml:mi>k</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>K</mml:mi></mml:mrow></mml:msubsup><mml:mrow><mml:msup><mml:mrow><mml:mfenced separators="|"><mml:mrow><mml:msub><mml:mrow><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>β</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:mo>-</mml:mo><mml:mi>β</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mfenced></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup></mml:mrow></mml:mrow></mml:mrow><mml:mrow><mml:mi>K</mml:mi></mml:mrow></mml:mfrac></mml:msqrt><mml:mi> </mml:mi><mml:mo>#</mml:mo><mml:mfenced separators="|"><mml:mrow><mml:mn>8</mml:mn></mml:mrow></mml:mfenced></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
	<p>where <inline-formula><mml:math id="m179"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>β</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> is the estimated regression coefficient in cluster <inline-formula><mml:math id="m180"><mml:mi>k</mml:mi><mml:mi> </mml:mi></mml:math></inline-formula>and <inline-formula><mml:math id="m181"><mml:msub><mml:mrow><mml:mi>β</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> is the corresponding true value. To get a better idea of the separate <inline-formula><mml:math id="m182"><mml:mi>β</mml:mi></mml:math></inline-formula> coefficients in different clusters (i.e., without averaging across clusters), we also reported the bias for <inline-formula><mml:math id="m183"><mml:msub><mml:mrow><mml:mi>β</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:math></inline-formula> in the Supplementary Material as an example (S2; <xref ref-type="bibr" rid="r44">Zhao et al., 2025b</xref>), with positive values indicating overestimation in most conditions. Similar trends were observed for <inline-formula><mml:math id="m184"><mml:msub><mml:mrow><mml:mi>β</mml:mi></mml:mrow><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msub></mml:math></inline-formula> and <inline-formula><mml:math id="m185"><mml:msub><mml:mrow><mml:mi>β</mml:mi></mml:mrow><mml:mrow><mml:mn>4</mml:mn></mml:mrow></mml:msub></mml:math></inline-formula>. For <inline-formula><mml:math id="m186"><mml:msub><mml:mrow><mml:mi>β</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:math></inline-formula>, the bias values were the smallest and close to zero (slightly negative).</p><?figure f5?>
<p>On average, <inline-formula><mml:math id="m187"><mml:mi>M</mml:mi><mml:msub><mml:mrow><mml:mi>E</mml:mi></mml:mrow><mml:mrow><mml:mi>β</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> was 0.037, -0.004, 0.039, and 0.031 (<xref ref-type="table" rid="t3">Table 3</xref>), and <inline-formula><mml:math id="m188"><mml:mi>R</mml:mi><mml:mi>M</mml:mi><mml:mi>S</mml:mi><mml:msub><mml:mrow><mml:mi>E</mml:mi></mml:mrow><mml:mrow><mml:mi>β</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> was 0.073, 0.054, 0.072 and 0.069 (<xref ref-type="table" rid="t4">Table 4</xref>) for <inline-formula><mml:math id="m189"><mml:msub><mml:mrow><mml:mi>β</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:math></inline-formula>, <inline-formula><mml:math id="m190"><mml:msub><mml:mrow><mml:mi>β</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:math></inline-formula>, <inline-formula><mml:math id="m191"><mml:msub><mml:mrow><mml:mi>β</mml:mi></mml:mrow><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msub></mml:math></inline-formula>, and <inline-formula><mml:math id="m192"><mml:msub><mml:mrow><mml:mi>β</mml:mi></mml:mrow><mml:mrow><mml:mn>4</mml:mn></mml:mrow></mml:msub></mml:math></inline-formula>, respectively.<xref ref-type="fn" rid="fn7"><sup>7</sup></xref><fn id="fn7"><label>7</label>
		<p>We also checked the distributions of <inline-formula><mml:math id="m193"><mml:mi>M</mml:mi><mml:msub><mml:mrow><mml:mi>E</mml:mi></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>β</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:msub></mml:math></inline-formula>, which were mostly symmetric (see <xref ref-type="fig" rid="f2">Figure 2</xref> in the Supplementary Material, S3; <xref ref-type="bibr" rid="r44">Zhao et al., 2025b</xref>). Thus, we retained the use of means and <italic>SD</italic>s and did not include a separate table for the medians and MADs for these outcomes.</p></fn> We performed an ANOVA on <inline-formula><mml:math id="m194"><mml:mi>M</mml:mi><mml:msub><mml:mrow><mml:mi>E</mml:mi></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>β</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:msub></mml:math></inline-formula> as an example (S1). All main effects were significant at the <inline-formula><mml:math id="m195"><mml:mi>α</mml:mi><mml:mo>=</mml:mo><mml:mn>0.01</mml:mn></mml:math></inline-formula> level, except for AMI. Similar to the trends observed for the cluster recovery, larger <inline-formula><mml:math id="m196"><mml:mi>G</mml:mi></mml:math></inline-formula>, larger <inline-formula><mml:math id="m197"><mml:msub><mml:mrow><mml:mi>N</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, smaller <inline-formula><mml:math id="m198"><mml:mi>K</mml:mi></mml:math></inline-formula>, larger <inline-formula><mml:math id="m199"><mml:mi>β</mml:mi></mml:math></inline-formula>, higher reliability, and smaller crossloadings resulted in smaller <inline-formula><mml:math id="m200"><mml:mi>M</mml:mi><mml:msub><mml:mrow><mml:mi>E</mml:mi></mml:mrow><mml:mrow><mml:mi>β</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, with crossloadings having the strongest effect (<xref ref-type="table" rid="t3">Table 3</xref>). To get a better idea of how <inline-formula><mml:math id="m201"><mml:mi>M</mml:mi><mml:msub><mml:mrow><mml:mi>E</mml:mi></mml:mrow><mml:mrow><mml:mi>β</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> got influenced by the manipulated factors, we depicted the interaction effects of <inline-formula><mml:math id="m202"><mml:msub><mml:mrow><mml:mi>N</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, <inline-formula><mml:math id="m203"><mml:mi>K</mml:mi></mml:math></inline-formula>, and <inline-formula><mml:math id="m204"><mml:mi>β</mml:mi></mml:math></inline-formula> across different levels of crossloadings and reliability (<xref ref-type="fig" rid="f6">Figure 6</xref>). Specifically, larger crossloadings resulted in larger <inline-formula><mml:math id="m205"><mml:msub><mml:mrow><mml:mi>M</mml:mi><mml:mi>E</mml:mi></mml:mrow><mml:mrow><mml:mi>β</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> values, especially in case of lower item reliability. Note that the recovery of the regression parameters was barely affected by using different prior variances, even more narrow ones, likely due to the fact that the cluster recovery was hardly affected as well.</p>
<table-wrap id="t3" position="anchor" orientation="portrait">
<label>Table 3</label><caption><title>The Average <inline-formula><mml:math id="m206"><mml:mi mathvariant="normal">M</mml:mi><mml:msub><mml:mrow><mml:mi>E</mml:mi></mml:mrow><mml:mrow><mml:mi>β</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> for Each of the Four Estimated Regression Parameters When Using the True Prior Variances for the Loadings</title></caption>
<table frame="hsides" rules="groups">
<col width="" align="left"/>
<col width=""/>
<col width=""/>
<col width=""/>
<col width=""/>
<col width=""/>
<thead>
<tr>
<th>Factor</th>
<th>Level</th>
<th><inline-formula><mml:math id="m207"><mml:mi>M</mml:mi><mml:msub><mml:mrow><mml:mi>E</mml:mi></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>β</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:msub></mml:math></inline-formula> (SD)</th>
<th><inline-formula><mml:math id="m208"><mml:mi>M</mml:mi><mml:msub><mml:mrow><mml:mi>E</mml:mi></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>β</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:msub></mml:math></inline-formula> (SD)</th>
<th><inline-formula><mml:math id="m209"><mml:mi>M</mml:mi><mml:msub><mml:mrow><mml:mi>E</mml:mi></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>β</mml:mi></mml:mrow><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:msub></mml:math></inline-formula> (SD)</th>
<th><inline-formula><mml:math id="m210"><mml:mi>M</mml:mi><mml:msub><mml:mrow><mml:mi>E</mml:mi></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>β</mml:mi></mml:mrow><mml:mrow><mml:mn>4</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:msub></mml:math></inline-formula> (SD)</th>
</tr>
</thead>
<tbody>
<tr>
<td><inline-formula><mml:math id="m211"><mml:mi>G</mml:mi></mml:math></inline-formula></td>
<td>24</td>
<td>0.037 (0.057)</td>
<td>-0.004 (0.045)</td>
<td>0.039 (0.053)</td>
<td>0.032 (0.051)</td>
</tr>
<tr>
<td/>	
<td>48</td>
<td>0.036 (0.046)</td>
<td>-0.004 (0.033)</td>
<td>0.039 (0.043)</td>
<td>0.031 (0.041)</td>
</tr>
<tr style="grey-border-top">
<td><inline-formula><mml:math id="m212"><mml:msub><mml:mrow><mml:mi>N</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula></td>
<td>50</td>
<td>0.040 (0.071)</td>
<td>-0.002 (0.059)</td>
<td>0.040 (0.063)</td>
<td>0.034 (0.063)</td>
</tr>
<tr>
<td/>	
<td>100</td>
<td>0.035 (0.044)</td>
<td>-0.005 (0.031)</td>
<td>0.038 (0.043)</td>
<td>0.030 (0.040)</td>
</tr>
<tr>
<td/>	
<td>200</td>
<td>0.034 (0.033)</td>
<td>-0.006 (0.018)</td>
<td>0.038 (0.036)</td>
<td>0.029 (0.031)</td>
</tr>
<tr style="grey-border-top">
<td><inline-formula><mml:math id="m213"><mml:mi>K</mml:mi></mml:math></inline-formula></td>
<td>2</td>
<td>0.035 (0.046)</td>
<td>-0.005 (0.036)</td>
<td>0.038 (0.042)</td>
<td>0.031 (0.039)</td>
</tr>
<tr>
<td/>	
<td>4</td>
<td>0.038 (0.056)</td>
<td>-0.004 (0.043)</td>
<td>0.040 (0.054)</td>
<td>0.032 (0.053)</td>
</tr>
<tr style="grey-border-top">
<td><inline-formula><mml:math id="m214"><mml:mi>β</mml:mi></mml:math></inline-formula></td>
<td align="char" char=".">0.2</td>
<td>0.040 (0.059)</td>
<td>-0.002 (0.048)</td>
<td>0.041 (0.055)</td>
<td>0.036 (0.052)</td>
</tr>
<tr>
<td/>	
<td align="char" char=".">0.3</td>
<td>0.037 (0.050)</td>
<td>-0.004 (0.038)</td>
<td>0.039 (0.047)</td>
<td>0.032 (0.046)</td>
</tr>
<tr>
<td/>	
<td align="char" char=".">0.4</td>
<td>0.033 (0.044)</td>
<td>-0.007 (0.031)</td>
<td>0.037 (0.042)</td>
<td>0.026 (0.041)</td>
</tr>
<tr style="grey-border-top">
<td>Reliability</td>
<td>high</td>
<td>0.033 (0.045)</td>
<td>-0.005 (0.035)</td>
<td>0.035 (0.043)</td>
<td>0.028 (0.042)</td>
</tr>
<tr>
<td/>	
<td>low</td>
<td>0.040 (0.058)</td>
<td>-0.004 (0.044)</td>
<td>0.042 (0.053)</td>
<td>0.034 (0.051)</td>
</tr>
<tr style="grey-border-top">
<td>AMI</td>
<td align="char" char=".">0.001</td>
<td>0.037 (0.050)</td>
<td>-0.004 (0.039)</td>
<td>0.040 (0.048)</td>
<td>0.032 (0.046)</td>
</tr>
<tr>
<td/>	
<td align="char" char=".">0.005</td>
<td>0.036 (0.050)</td>
<td>-0.004 (0.039)</td>
<td>0.040 (0.048)</td>
<td>0.033 (0.046)</td>
</tr>
<tr>
<td/>	
<td align="char" char=".">0.01</td>
<td>0.037 (0.051)</td>
<td>-0.004 (0.039)</td>
<td>0.039 (0.048)</td>
<td>0.032 (0.046)</td>
</tr>
<tr>
<td/>	
<td align="char" char=".">0.05</td>
<td>0.037 (0.053)</td>
<td>-0.005 (0.041)</td>
<td>0.039 (0.050)</td>
<td>0.030 (0.047)</td>
</tr>
<tr>
<td/>	
<td align="char" char=".">0.1</td>
<td>0.036 (0.054)</td>
<td>-0.005 (0.041)</td>
<td>0.037 (0.048)</td>
<td>0.029 (0.048)</td>
</tr>
<tr style="grey-border-top">
<td>Crossloadings</td>
<td>0</td>
<td>0.001 (0.043)</td>
<td>0.001 (0.040)</td>
<td>-0.000 (0.036)</td>
<td>0.002 (0.040)</td>
</tr>
<tr>
<td/>	
<td align="char" char=".">0.2</td>
<td>0.040 (0.043)</td>
<td>-0.004 (0.040)</td>
<td>0.042 (0.037)</td>
<td>0.033 (0.040)</td>
</tr>
<tr>
<td/>	
<td align="char" char=".">0.4</td>
<td>0.069 (0.045)</td>
<td>-0.010 (0.039)</td>
<td>0.075 (0.039)</td>
<td>0.059 (0.041)</td>
</tr>
<tr style="background-lightblue; white-border-top; white-border-bottom">
<td align="left" colspan="2">Total</td>
<td>0.037 (0.052)</td>
<td>-0.004 (0.040)</td>
<td>0.039 (0.048)</td>
<td>0.031 (0.047)</td>
</tr>
</tbody>
</table>
</table-wrap>
<table-wrap id="t4" position="anchor" orientation="portrait">
<?pagebreak-before?><label>Table 4</label><caption><title>The Average <inline-formula><mml:math id="m215"><mml:mi mathvariant="normal">R</mml:mi><mml:mi mathvariant="normal">M</mml:mi><mml:mi mathvariant="normal">S</mml:mi><mml:msub><mml:mrow><mml:mi>E</mml:mi></mml:mrow><mml:mrow><mml:mi>β</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> for Each of the Four Estimated Regression Parameters When Using the True Prior Variances for the Loadings</title></caption>
<table frame="hsides" rules="groups">
<col width="" align="left"/>
<col width=""/>
<col width=""/>
<col width=""/>
<col width=""/>
<col width=""/>
<thead>
<tr>
<th>Factor</th>
<th>Level</th>
<th><inline-formula><mml:math id="m216"><mml:msub><mml:mrow><mml:mi>R</mml:mi><mml:mi>M</mml:mi><mml:mi>S</mml:mi><mml:mi>E</mml:mi></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>β</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:msub></mml:math></inline-formula> (SD)</th>
<th><inline-formula><mml:math id="m217"><mml:msub><mml:mrow><mml:mi>R</mml:mi><mml:mi>M</mml:mi><mml:mi>S</mml:mi><mml:mi>E</mml:mi></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>β</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:msub></mml:math></inline-formula> (SD)</th>
<th><inline-formula><mml:math id="m218"><mml:msub><mml:mrow><mml:mi>R</mml:mi><mml:mi>M</mml:mi><mml:mi>S</mml:mi><mml:mi>E</mml:mi></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>β</mml:mi></mml:mrow><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:msub></mml:math></inline-formula> (SD)</th>
<th><inline-formula><mml:math id="m219"><mml:msub><mml:mrow><mml:mi>R</mml:mi><mml:mi>M</mml:mi><mml:mi>S</mml:mi><mml:mi>E</mml:mi></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>β</mml:mi></mml:mrow><mml:mrow><mml:mn>4</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:msub></mml:math></inline-formula> (SD)</th>
</tr>
</thead>
<tbody>
<tr>
<td><inline-formula><mml:math id="m220"><mml:mi>G</mml:mi></mml:math></inline-formula></td>
<td>24</td>
<td>0.081 (0.065)</td>
<td>0.062 (0.057)</td>
<td>0.079 (0.058)</td>
<td>0.077 (0.061)</td>
</tr>
<tr>
<td/>	
<td>48</td>
<td>0.064 (0.051)</td>
<td>0.045 (0.044)</td>
<td>0.065 (0.046)</td>
<td>0.061 (0.048)</td>
</tr>
<tr style="grey-border-top">
<td><inline-formula><mml:math id="m221"><mml:msub><mml:mrow><mml:mi>N</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula></td>
<td>50</td>
<td>0.105 (0.079)</td>
<td>0.087 (0.069)</td>
<td>0.100 (0.070)</td>
<td>0.101 (0.075)</td>
</tr>
<tr>
<td/>	
<td>100</td>
<td>0.064 (0.042)</td>
<td>0.047 (0.035)</td>
<td>0.065 (0.039)</td>
<td>0.061 (0.039)</td>
</tr>
<tr>
<td/>	
<td>200</td>
<td>0.048 (0.027)</td>
<td>0.027 (0.017)</td>
<td>0.051 (0.029)</td>
<td>0.045 (0.024)</td>
</tr>
<tr style="grey-border-top">
<td><inline-formula><mml:math id="m222"><mml:mi>K</mml:mi></mml:math></inline-formula></td>
<td>2</td>
<td>0.055 (0.039)</td>
<td>0.036 (0.032)</td>
<td>0.056 (0.037)</td>
<td>0.052 (0.035)</td>
</tr>
<tr>
<td/>	
<td>4</td>
<td>0.090 (0.070)</td>
<td>0.071 (0.061)</td>
<td>0.088 (0.061)</td>
<td>0.087 (0.066)</td>
</tr>
<tr style="grey-border-top">
<td><inline-formula><mml:math id="m223"><mml:mi>β</mml:mi></mml:math></inline-formula></td>
<td align="char" char=".">0.2</td>
<td>0.086 (0.067)</td>
<td>0.067 (0.061)</td>
<td>0.085 (0.062)</td>
<td>0.083 (0.062)</td>
</tr>
<tr>
<td/>	
<td align="char" char=".">0.3</td>
<td>0.070 (0.057)</td>
<td>0.051 (0.050)</td>
<td>0.070 (0.051)</td>
<td>0.067 (0.054)</td>
</tr>
<tr>
<td/>	
<td align="char" char=".">0.4</td>
<td>0.061 (0.049)</td>
<td>0.043 (0.038)</td>
<td>0.060 (0.041)</td>
<td>0.058 (0.048)</td>
</tr>
<tr style="grey-border-top">
<td>Reliability</td>
<td>high</td>
<td>0.064 (0.048)</td>
<td>0.048 (0.045)</td>
<td>0.065 (0.047)</td>
<td>0.062 (0.048)</td>
</tr>
<tr>
<td/>	
<td>low</td>
<td>0.081 (0.068)</td>
<td>0.059 (0.057)</td>
<td>0.079 (0.058)</td>
<td>0.076 (0.062)</td>
</tr>
<tr style="grey-border-top">
<td>AMI</td>
<td align="char" char=".">0.001</td>
<td>0.072 (0.055)</td>
<td>0.052 (0.050)</td>
<td>0.071 (0.050)</td>
<td>0.068 (0.053)</td>
</tr>
<tr>
<td/>	
<td align="char" char=".">0.005</td>
<td>0.072 (0.056)</td>
<td>0.053 (0.051)</td>
<td>0.071 (0.052)</td>
<td>0.069 (0.054)</td>
</tr>
<tr>
<td/>	
<td align="char" char=".">0.01</td>
<td>0.072 (0.057)</td>
<td>0.053 (0.051)</td>
<td>0.071 (0.051)</td>
<td>0.069 (0.054)</td>
</tr>
<tr>
<td/>	
<td align="char" char=".">0.05</td>
<td>0.074 (0.062)</td>
<td>0.055 (0.052)</td>
<td>0.073 (0.057)</td>
<td>0.070 (0.057)</td>
</tr>
<tr>
<td/>	
<td align="char" char=".">0.1</td>
<td>0.074 (0.066)</td>
<td>0.055 (0.054)</td>
<td>0.072 (0.054)</td>
<td>0.070 (0.061)</td>
</tr>
<tr style="grey-border-top">
<td>Crossloadings</td>
<td>0</td>
<td>0.055 (0.058)</td>
<td>0.052 (0.052)</td>
<td>0.051 (0.052)</td>
<td>0.056 (0.055)</td>
</tr>
<tr>
<td/>	
<td align="char" char=".">0.2</td>
<td>0.070 (0.058)</td>
<td>0.053 (0.053)</td>
<td>0.069 (0.050)</td>
<td>0.067 (0.055)</td>
</tr>
<tr>
<td/>	
<td align="char" char=".">0.4</td>
<td>0.092 (0.056)</td>
<td>0.055 (0.050)</td>
<td>0.096 (0.048)</td>
<td>0.085 (0.053)</td>
</tr>
<tr style="background-lightblue; white-border-top; white-border-bottom">
<td align="left" colspan="2">Total</td>
<td>0.073 (0.059)</td>
<td>0.054 (0.052)</td>
<td>0.072 (0.053)</td>
<td>0.069 (0.056)</td>
</tr>
</tbody>
</table>
</table-wrap><fig id="f6" position="anchor" fig-type="figure" orientation="portrait"><label>Figure 6</label><caption>
<title>The <inline-formula><mml:math id="m224"><mml:mi>M</mml:mi><mml:msub><mml:mrow><mml:mi>E</mml:mi></mml:mrow><mml:mrow><mml:mi>β</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> for MixMG-BSEM in Function of the Within-Group Sample Sizes, Number of Clusters, Size of Regression Parameters, Crossloadings and Reliability</title></caption><graphic xlink:href="meth.16411-f6" position="anchor" orientation="portrait"/></fig></sec>
<sec sec-type="conclusions"><title>Section Conclusion</title>
<p>We assessed the performance of MixMG-BSEM when the true number of clusters was known. We found that performing 50 random starts in Step 3 largely prevented local maxima. The recovery of clusters and regression parameters was good when the within-group sample size was at least 200 and/or in case of a larger cluster separation (i.e., <inline-formula><mml:math id="m225"><mml:mi>β</mml:mi><mml:mo>=</mml:mo><mml:mn>0.4</mml:mn></mml:math></inline-formula>). Ignoring larger crossloadings (by estimating the MM per factor) resulted in more biased estimates for factor loadings and regression parameters, while cluster recovery was less affected. Although prior selection based on the DIC was not always optimal, the recovery of clusters and regression parameters was relatively stable across the priors.</p></sec></sec>
	<sec sec-type="discussion"><title>Discussion</title><?pagebreak-before?>
<p>We presented MixMG-BSEM as a new addition to the novel mixture SEM framework for comparing structural relations across many groups. Unlike the existing approaches that rely on the exact MI assumption, MixMG-BSEM adopts the more realistic assumption of AMI, which accommodates small differences in MM parameters across groups. Specifically, after estimating the MM using MG-BCFA with small-variance priors, MixMG-BSEM clusters groups with the same structural relations, thereby eliminating the need for pairwise comparisons of group-specific structural relations. Since its results may depend on the specified prior variances, results obtained with different prior variances should be compared and the best prior selected based on model fit measures such as the Deviance Information Criterion (DIC). Our simulation study results showed that both the clustering and regression parameter estimates were relatively insensitive to the choice of prior variances, however.</p>
<p>Currently, MixMG-BSEM estimates the MM per factor (i.e., with one factor per measurement block). In the simulation study, the cluster recovery was unaffected by ignoring crossloadings, but the recovery of the factor loadings and regression estimates was affected. Therefore, it would be valuable to investigate the performance of MixMG-BSEM when including factors with crossloadings in the same measurement block, at the cost of a (much) longer computation time, where small-variance priors could also be applied to the crossloadings to allow for small differences (<xref ref-type="bibr" rid="r25">Muthén &amp; Asparouhov, 2012</xref>). However, it is important to note that the default prior mean for crossloadings is zero, whereas applying a prior mean of zero to a sizeable crossloading can negatively impact the regression parameter estimates (<xref ref-type="bibr" rid="r40">Wei et al., 2022</xref>). Therefore, researchers should gather prior information about crossloadings before choosing an appropriate prior (<xref ref-type="bibr" rid="r40">Wei et al., 2022</xref>). Note that, in cases with several crossloadings connecting more than two factors, the estimation may fail when all these factors are included in one measurement block, especially in case of many groups. Thus, in such cases, model estimation is only possible when partitioning the factors into smaller measurement blocks, highlighting the benefits of the measurement block approach even more.</p>
<p>While the simulation study evaluated the performance of MixMG-BSEM with approximate metric invariance for all loadings, except for the invariant marker variable loading, MixMG-BSEM can theoretically accommodate all combinations of exact, approximate and non-invariance for the loadings. The stepwise estimation of MixMG-BSEM conveniently allows to tweak the MG-BCFA model, for instance, by specifying certain loadings as non-invariant, before moving onto the next steps. Similarly, if group-specific loading estimates are virtually identical across groups, one may consider specifying the loading as exactly invariant. Specifying an invariant parameter as approximately invariant is rather harmless, whereas specifying a non-invariant parameter as approximately invariant may introduce bias in parameter estimation and affect the clustering. Note that MG-BCFA allows to evaluate non-invariances for all parameters, which is achieved by comparing group-specific estimates to the credible intervals of the average posterior estimates across all groups (e.g., <xref ref-type="bibr" rid="r41">Winter &amp; Depaoli, 2020</xref>). In future research, it would be interesting to evaluate the performance of MixMG-BSEM when non-invariant loadings are specified as approximately invariant.</p>
<p>The simulation study assumed the number of clusters to be known, whereas this is typically unknown for empirical data. To determine the number of clusters, different methods are available, such as the Bayesian Information Criterion (BIC; <xref ref-type="bibr" rid="r34">Schwarz, 1978</xref>), Akaike Information Criterion (AIC; <xref ref-type="bibr" rid="r1">Akaike, 1974</xref>), and convex hull procedure (CHull; <xref ref-type="bibr" rid="r7">Ceulemans &amp; Kiers, 2006</xref>). In brief, all these methods balance model fit (i.e., the log-likelihood) and model complexity (i.e., the number of parameters). BIC and AIC do so by combining model fit and a penalty for model complexity into a single criterion, whereas CHull uses a generalized scree test. Previous studies on model selection for MixMG-SEM (<xref ref-type="bibr" rid="r30">Perez Alonso et al., 2025</xref>) and MixML-SEM (<xref ref-type="bibr" rid="r43">Zhao et al., 2025a</xref>) have shown that combining AIC, BIC, and CHull — with visual inspection of the scree plot — is an effective way to determine the number of clusters. Since MixMG-BSEM performs the same mixture clustering on group-specific factor covariances as these methods, we expect these recommendations to generalize to MixMG-BSEM. However, in the future, it would still be useful to evaluate model selection for MixMG-BSEM specifically.</p>
<p>Currently, MixMG-BSEM combines Bayesian and maximum likelihood estimation, assuming continuous items. In empirical practice, we often work with ordinal items with a few response categories (e.g., Likert scale items). To accommodate ordinal data in MixMG-BSEM, only the first step (i.e., MG-BCFA) would need to be adjusted to deal with ordinal data (<xref ref-type="bibr" rid="r26">Muthén &amp; Asparouhov, 2013a</xref>), whereas the subsequent steps would remain unchanged. In future studies, it will be valuable to evaluate the performance of MixMG-BSEM adapted to ordinal data.</p>
<p>In conclusion, MixMG-BSEM is an effective method for accommodating AMI while clustering structural relations of interest. By relaxing the strict assumption of exact MI, it extends the framework of novel mixture SEM methods in an important way, making it more suited for empirical applications where small differences in parameters across groups are expected.</p>
</sec>
</body>
<back><fn-group><fn fn-type="financial-disclosure">
<p content-type="fn-title">Funded by the European Union (ERC, PROCESSHETEROGENEITY, 101040754, awarded to Kim De Roover). Views and opinions expressed are however those of the author(s) only and do not necessarily reflect those of the European Union or European Research Council Executive Agency. Neither the European Union nor the granting authority can be held responsible for them. The resources and services used in this work were provided by the VSC (Flemish Supercomputer Center), funded by the Research Foundation — Flanders (FWO) and the Flemish Government.</p></fn></fn-group>
<ref-list><title>References</title>
<ref id="r1"><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name name-style="western"><surname>Akaike</surname>, <given-names>H.</given-names></string-name></person-group> (<year>1974</year>). <article-title>A new look at the statistical model identification.</article-title> <source>IEEE Transactions on Automatic Control</source>, <volume>19</volume>(<issue>6</issue>), <fpage>716</fpage>–<lpage>723</lpage>. <pub-id pub-id-type="doi">10.1109/TAC.1974.1100705</pub-id></mixed-citation></ref>
<ref id="r2"><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name name-style="western"><surname>Arminger</surname>, <given-names>G.</given-names></string-name>, &amp; <string-name name-style="western"><surname>Stein</surname>, <given-names>P.</given-names></string-name></person-group> (<year>1997</year>). <article-title>Finite mixtures of covariance structure models with regressors: Loglikelihood function, minimum distance estimation, fit indices, and a complex example.</article-title> <source>Sociological Methods &amp; Research</source>, <volume>26</volume>(<issue>2</issue>), <fpage>148</fpage>–<lpage>182</lpage>. <pub-id pub-id-type="doi">10.1177/0049124197026002002</pub-id></mixed-citation></ref>
<ref id="r3"><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name name-style="western"><surname>Arts</surname>, <given-names>I.</given-names></string-name>, <string-name name-style="western"><surname>Fang</surname>, <given-names>Q.</given-names></string-name>, <string-name name-style="western"><surname>van de Schoot</surname>, <given-names>R.</given-names></string-name>, &amp; <string-name name-style="western"><surname>Meitinger</surname>, <given-names>K.</given-names></string-name></person-group> (<year>2021</year>). <article-title>Approximate measurement invariance of willingness to sacrifice for the environment across 30 countries: The importance of prior distributions and their visualization.</article-title> <source>Frontiers in Psychology</source>, <volume>12</volume>, <elocation-id>624032</elocation-id>. <pub-id pub-id-type="doi">10.3389/fpsyg.2021.624032</pub-id><pub-id pub-id-type="pmid">34366953</pub-id></mixed-citation></ref>
<ref id="r4"><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name name-style="western"><surname>Asparouhov</surname>, <given-names>T.</given-names></string-name>, <string-name name-style="western"><surname>Muthén</surname>, <given-names>B. O.</given-names></string-name>, &amp; <string-name name-style="western"><surname>Morin</surname>, <given-names>A. J. S.</given-names></string-name></person-group> (<year>2015</year>). <article-title>Bayesian structural equation modeling with cross-loadings and residual covariances: Comments on Stromeyer et al.</article-title> <source>Journal of Management</source>, <volume>41</volume>(<issue>6</issue>), <fpage>1561</fpage>–<lpage>1577</lpage>. <pub-id pub-id-type="doi">10.1177/0149206315591075</pub-id></mixed-citation></ref>
<ref id="r5"><mixed-citation publication-type="book">Bollen, K. A. (1989). <italic>Structural equations with latent variables</italic>. John Wiley &amp; Sons. <pub-id pub-id-type="doi">10.1002/9781118619179</pub-id></mixed-citation></ref>
<ref id="r6"><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name name-style="western"><surname>Byrne</surname>, <given-names>B. M.</given-names></string-name>, <string-name name-style="western"><surname>Shavelson</surname>, <given-names>R. J.</given-names></string-name>, &amp; <string-name name-style="western"><surname>Muthén</surname>, <given-names>B. O.</given-names></string-name></person-group> (<year>1989</year>). <article-title>Testing for the equivalence of factor covariance and mean structures: The issue of partial measurement invariance.</article-title> <source>Psychological Bulletin</source>, <volume>105</volume>(<issue>3</issue>), <fpage>456</fpage>–<lpage>466</lpage>. <pub-id pub-id-type="doi">10.1037/0033-2909.105.3.456</pub-id></mixed-citation></ref>
<ref id="r7"><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name name-style="western"><surname>Ceulemans</surname>, <given-names>E.</given-names></string-name>, &amp; <string-name name-style="western"><surname>Kiers</surname>, <given-names>H. A. L.</given-names></string-name></person-group> (<year>2006</year>). <article-title>Selecting among three-mode principal component models of different types and complexities: A numerical convex hull based method.</article-title> <source>British Journal of Mathematical &amp; Statistical Psychology</source>, <volume>59</volume>(<issue>1</issue>), <fpage>133</fpage>–<lpage>150</lpage>. <pub-id pub-id-type="doi">10.1348/000711005X64817</pub-id><pub-id pub-id-type="pmid">16709283</pub-id></mixed-citation></ref>
<ref id="r8"><mixed-citation publication-type="book">Croon, M. (2002). Using predicted latent scores in general latent structure models. In G. A. Marcoulides &amp; I. Moustaki (Eds.), <italic>Latent variable and latent structure models</italic> (1<sup>st</sup> ed., pp. 195–223). Lawrence Erlbaum.</mixed-citation></ref>
<ref id="r9"><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name name-style="western"><surname>Dempster</surname>, <given-names>A. P.</given-names></string-name>, <string-name name-style="western"><surname>Laird</surname>, <given-names>N. M.</given-names></string-name>, &amp; <string-name name-style="western"><surname>Rubin</surname>, <given-names>D. B.</given-names></string-name></person-group> (<year>1977</year>). <article-title>Maximum likelihood from incomplete data via the EM algorithm.</article-title> <source>Journal of the Royal Statistical Society: Series B (Methodological)</source>, <volume>39</volume>(<issue>1</issue>), <fpage>1</fpage>–<lpage>22</lpage>. <pub-id pub-id-type="doi">10.1111/j.2517-6161.1977.tb01600.x</pub-id></mixed-citation></ref>
<ref id="r10"><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name name-style="western"><surname>Dolan</surname>, <given-names>C. V.</given-names></string-name>, &amp; <string-name name-style="western"><surname>van der Maas</surname>, <given-names>H. L. J.</given-names></string-name></person-group> (<year>1998</year>). <article-title>Fitting multivariage normal finite mixtures subject to structural equation modeling.</article-title> <source>Psychometrika</source>, <volume>63</volume>(<issue>3</issue>), <fpage>227</fpage>–<lpage>253</lpage>. <pub-id pub-id-type="doi">10.1007/BF02294853</pub-id></mixed-citation></ref>
<ref id="r11"><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name name-style="western"><surname>Gelman</surname>, <given-names>A.</given-names></string-name>, <string-name name-style="western"><surname>Meng</surname>, <given-names>X.-L.</given-names></string-name>, &amp; <string-name name-style="western"><surname>Stern</surname>, <given-names>H.</given-names></string-name></person-group> (<year>1996</year>). <article-title>Posterior predictive assessment of model fitness via realized discrepancies.</article-title> <source>Statistica Sinica</source>, <volume>6</volume>(<issue>4</issue>), <fpage>733</fpage>–<lpage>760</lpage>.</mixed-citation></ref>
<ref id="r12"><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name name-style="western"><surname>Hoofs</surname>, <given-names>H.</given-names></string-name>, <string-name name-style="western"><surname>Van De Schoot</surname>, <given-names>R.</given-names></string-name>, <string-name name-style="western"><surname>Jansen</surname>, <given-names>N. W. H.</given-names></string-name>, &amp; <string-name name-style="western"><surname>Kant</surname>, <given-names>I.</given-names></string-name></person-group> (<year>2018</year>). <article-title>Evaluating model fit in Bayesian confirmatory factor analysis with large samples: Simulation study introducing the BRMSEA.</article-title> <source>Educational and Psychological Measurement</source>, <volume>78</volume>(<issue>4</issue>), <fpage>537</fpage>–<lpage>568</lpage>. <pub-id pub-id-type="doi">10.1177/0013164417709314</pub-id><pub-id pub-id-type="pmid">30034027</pub-id></mixed-citation></ref>
<ref id="r13"><mixed-citation publication-type="book">Hoyle, R. H. (2012). <italic>Handbook of structural equation modeling</italic>. Guilford Press.</mixed-citation></ref>
<ref id="r14"><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name name-style="western"><surname>Hubert</surname>, <given-names>L.</given-names></string-name>, &amp; <string-name name-style="western"><surname>Arabie</surname>, <given-names>P.</given-names></string-name></person-group> (<year>1985</year>). <article-title>Comparing partitions.</article-title> <source>Journal of Classification</source>, <volume>2</volume>(<issue>1</issue>), <fpage>193</fpage>–<lpage>218</lpage>. <pub-id pub-id-type="doi">10.1007/BF01908075</pub-id></mixed-citation></ref>
<ref id="r15"><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name name-style="western"><surname>Jedidi</surname>, <given-names>K.</given-names></string-name>, <string-name name-style="western"><surname>Jagpal</surname>, <given-names>H. S.</given-names></string-name>, &amp; <string-name name-style="western"><surname>Desarbo</surname>, <given-names>W. S.</given-names></string-name></person-group> (<year>1997</year>). <article-title>Finite-mixture structural equation models for response-based segmentation and unobserved heterogeneity.</article-title> <source>Marketing Science</source>, <volume>16</volume>(<issue>1</issue>), <fpage>39</fpage>–<lpage>59</lpage>. <pub-id pub-id-type="doi">10.1287/mksc.16.1.39</pub-id></mixed-citation></ref>
<ref id="r16"><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name name-style="western"><surname>Kim</surname>, <given-names>E. S.</given-names></string-name>, <string-name name-style="western"><surname>Cao</surname>, <given-names>C.</given-names></string-name>, <string-name name-style="western"><surname>Wang</surname>, <given-names>Y.</given-names></string-name>, &amp; <string-name name-style="western"><surname>Nguyen</surname>, <given-names>D. T.</given-names></string-name></person-group> (<year>2017</year>). <article-title>Measurement invariance testing with many groups: A comparison of five approaches.</article-title> <source>Structural Equation Modeling</source>, <volume>24</volume>(<issue>4</issue>), <fpage>524</fpage>–<lpage>544</lpage>. <pub-id pub-id-type="doi">10.1080/10705511.2017.1304822</pub-id></mixed-citation></ref>

<ref id="r18"><mixed-citation publication-type="book">Lek, K., Oberski, D., Davidov, E., Cieciuch, J., Seddig, D., &amp; Schmidt, P. (2018). Approximate measurement invariance. In T. P. Johnson, B. Pennell, I. A. L. Stoop &amp; B. Dorer (Eds.), <italic>Advances in comparative survey methods</italic> (pp. 911–929). John Wiley &amp; Sons. <pub-id pub-id-type="doi">10.1002/9781118884997.ch41</pub-id></mixed-citation></ref>
<ref id="r19"><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name name-style="western"><surname>Marsh</surname>, <given-names>H. W.</given-names></string-name>, <string-name name-style="western"><surname>Lüdtke</surname>, <given-names>O.</given-names></string-name>, <string-name name-style="western"><surname>Muthén</surname>, <given-names>B. O.</given-names></string-name>, <string-name name-style="western"><surname>Asparouhov</surname>, <given-names>T.</given-names></string-name>, <string-name name-style="western"><surname>Morin</surname>, <given-names>A. J. S.</given-names></string-name>, <string-name name-style="western"><surname>Trautwein</surname>, <given-names>U.</given-names></string-name>, &amp; <string-name name-style="western"><surname>Nagengast</surname>, <given-names>B.</given-names></string-name></person-group> (<year>2010</year>). <article-title>A new look at the big five factor structure through exploratory structural equation modeling.</article-title> <source>Psychological Assessment</source>, <volume>22</volume>(<issue>3</issue>), <fpage>471</fpage>–<lpage>491</lpage>. <pub-id pub-id-type="doi">10.1037/a0019227</pub-id><pub-id pub-id-type="pmid">20822261</pub-id></mixed-citation></ref>
<ref id="r20"><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name name-style="western"><surname>Marsh</surname>, <given-names>H. W.</given-names></string-name>, <string-name name-style="western"><surname>Morin</surname>, <given-names>A. J. S.</given-names></string-name>, <string-name name-style="western"><surname>Parker</surname>, <given-names>P. D.</given-names></string-name>, &amp; <string-name name-style="western"><surname>Kaur</surname>, <given-names>G.</given-names></string-name></person-group> (<year>2014</year>). <article-title>Exploratory structural equation modeling: An integration of the best features of exploratory and confirmatory factor analysis.</article-title> <source>Annual Review of Clinical Psychology</source>, <volume>10</volume>, <fpage>85</fpage>–<lpage>110</lpage>. <pub-id pub-id-type="doi">10.1146/annurev-clinpsy-032813-153700</pub-id><pub-id pub-id-type="pmid">24313568</pub-id></mixed-citation></ref>
<ref id="r21"><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name name-style="western"><surname>Marsh</surname>, <given-names>H. W.</given-names></string-name>, <string-name name-style="western"><surname>Muthén</surname>, <given-names>B. O.</given-names></string-name>, <string-name name-style="western"><surname>Asparouhov</surname>, <given-names>T.</given-names></string-name>, <string-name name-style="western"><surname>Lüdtke</surname>, <given-names>O.</given-names></string-name>, <string-name name-style="western"><surname>Robitzsch</surname>, <given-names>A.</given-names></string-name>, <string-name name-style="western"><surname>Morin</surname>, <given-names>A. J. S.</given-names></string-name>, &amp; <string-name name-style="western"><surname>Trautwein</surname>, <given-names>U.</given-names></string-name></person-group> (<year>2009</year>). <article-title>Exploratory structural equation modeling, integrating CFA and EFA: Application to students’ evaluations of university teaching.</article-title> <source>Structural Equation Modeling</source>, <volume>16</volume>(<issue>3</issue>), <fpage>439</fpage>–<lpage>476</lpage>. <pub-id pub-id-type="doi">10.1080/10705510903008220</pub-id></mixed-citation></ref>
<ref id="r22"><mixed-citation publication-type="book">McLachlan, G., &amp; Peel, D. (2000). <italic>Finite mixture models</italic>. Wiley. <pub-id pub-id-type="doi">10.1002/0471721182</pub-id></mixed-citation></ref>
<ref id="r23"><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name name-style="western"><surname>Merkle</surname>, <given-names>E. C.</given-names></string-name>, <string-name name-style="western"><surname>Fitzsimmons</surname>, <given-names>E.</given-names></string-name>, <string-name name-style="western"><surname>Uanhoro</surname>, <given-names>J.</given-names></string-name>, &amp; <string-name name-style="western"><surname>Goodrich</surname>, <given-names>B.</given-names></string-name></person-group> (<year>2021</year>). <article-title>Efficient Bayesian structural equation modeling in Stan.</article-title> <source>Journal of Statistical Software</source>, <volume>100</volume>(<issue>6</issue>), <fpage>1</fpage>–<lpage>22</lpage>. <pub-id pub-id-type="doi">10.18637/jss.v100.i06</pub-id></mixed-citation></ref>
<ref id="r24"><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name name-style="western"><surname>Michael</surname>, <given-names>D.</given-names></string-name>, &amp; <string-name name-style="western"><surname>Kyriakides</surname>, <given-names>L.</given-names></string-name></person-group> (<year>2023</year>). <article-title>Mediating effects of motivation and socioeconomic status on reading achievement: A secondary analysis of PISA 2018.</article-title> <source>Large-Scale Assessments in Education</source>, <volume>11</volume>, <elocation-id>31</elocation-id>. <pub-id pub-id-type="doi">10.1186/s40536-023-00181-9</pub-id></mixed-citation></ref>
<ref id="r25"><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name name-style="western"><surname>Muthén</surname>, <given-names>B.</given-names></string-name>, &amp; <string-name name-style="western"><surname>Asparouhov</surname>, <given-names>T.</given-names></string-name></person-group> (<year>2012</year>). <article-title>Bayesian structural equation modeling: A more flexible representation of substantive theory.</article-title> <source>Psychological Methods</source>, <volume>17</volume>(<issue>3</issue>), <fpage>313</fpage>–<lpage>335</lpage>. <pub-id pub-id-type="doi">10.1037/a0026802</pub-id><pub-id pub-id-type="pmid">22962886</pub-id></mixed-citation></ref>
<ref id="r26"><mixed-citation publication-type="web">Muthén, B., &amp; Asparouhov, T. (2013a). <italic>BSEM measurement invariance analysis</italic> (Mplus Web Notes: No. 17). Mplus. <ext-link ext-link-type="uri" xlink:href="https://www.statmodel.com/examples/webnotes/webnote17.pdf">https://www.statmodel.com/examples/webnotes/webnote17.pdf</ext-link></mixed-citation></ref>
<ref id="r27"><mixed-citation publication-type="web">Muthén, B., &amp; Asparouhov, T. (2013b). <italic>New methods for the study of measurement invariance with many groups</italic>. Mplus. <ext-link ext-link-type="uri" xlink:href="https://www.statmodel.com/download/PolAn.pdf">https://www.statmodel.com/download/PolAn.pdf</ext-link></mixed-citation></ref>
<ref id="r28"><mixed-citation publication-type="book">Muthén, L. K., &amp; Muthén, B. O. (1998). <italic>Mplus user’s guide: Statistical analysis with latent variables</italic> (8<sup>th</sup> ed.). Muthén &amp; Muthén. <ext-link ext-link-type="uri" xlink:href="https://www.statmodel.com/download/usersguide/MplusUserGuideVer_8.pdf#:~:text=Following%20is%20the%20correct%20citation%20for%20this%20document%3A,Muth%C3%A9n%20%26%20Muth%C3%A9n%20Version%208%20April%202017%20">https://www.statmodel.com/download/usersguide/MplusUserGuideVer_8.pdf#:~:text=Following%20is%20the%20correct%20citation%20for%20this%20document%3A,Muth%C3%A9n%20%26%20Muth%C3%A9n%20Version%208%20April%202017%20</ext-link></mixed-citation></ref>
<ref id="r29"><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name name-style="western"><surname>Perez Alonso</surname>, <given-names>A. F.</given-names></string-name>, <string-name name-style="western"><surname>Rosseel</surname>, <given-names>Y.</given-names></string-name>, <string-name name-style="western"><surname>Vermunt</surname>, <given-names>J. K.</given-names></string-name>, &amp; <string-name name-style="western"><surname>De Roover</surname>, <given-names>K.</given-names></string-name></person-group> (<year>2024</year>). <article-title>Mixture multigroup structural equation modeling: A novel method for comparing structural relations across many groups.</article-title> <source>Psychological Methods</source>. <comment>Advance online publication</comment>. <pub-id pub-id-type="doi">10.1037/met0000667</pub-id><pub-id pub-id-type="pmid">39264645</pub-id></mixed-citation></ref>
<ref id="r30"><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name name-style="western"><surname>Perez Alonso</surname>, <given-names>A. F.</given-names></string-name>, <string-name name-style="western"><surname>Vermunt</surname>, <given-names>J. K.</given-names></string-name>, <string-name name-style="western"><surname>Rosseel</surname>, <given-names>Y.</given-names></string-name>, &amp; <string-name name-style="western"><surname>Roover</surname>, <given-names>K. D.</given-names></string-name></person-group> (<year>2025</year>). <article-title>Selecting the number of clusters in mixture multigroup structural equation modeling.</article-title> <source>Methodology: European Journal of Research Methods for the Behavioral and Social Sciences</source>, <volume>21</volume>(<issue>1</issue>), <elocation-id>e14931</elocation-id>. <pub-id pub-id-type="doi">10.5964/meth.14931</pub-id></mixed-citation></ref>
<ref id="r31"><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name name-style="western"><surname>Pokropek</surname>, <given-names>A.</given-names></string-name>, <string-name name-style="western"><surname>Schmidt</surname>, <given-names>P.</given-names></string-name>, &amp; <string-name name-style="western"><surname>Davidov</surname>, <given-names>E.</given-names></string-name></person-group> (<year>2020</year>). <article-title>Choosing priors in Bayesian measurement invariance modeling: A Monte Carlo simulation study.</article-title> <source>Structural Equation Modeling</source>, <volume>27</volume>(<issue>5</issue>), <fpage>750</fpage>–<lpage>764</lpage>. <pub-id pub-id-type="doi">10.1080/10705511.2019.1703708</pub-id></mixed-citation></ref>
	<ref id="r32"><mixed-citation publication-type="web">R Core Team. (2022). <italic>R: A language and environment for statistical computing</italic>. R Foundation for Statistical Computing. <ext-link ext-link-type="uri" xlink:href="https://www.R-project.org/">https://www.R-project.org/</ext-link></mixed-citation></ref>
<ref id="r33"><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name name-style="western"><surname>Rosseel</surname>, <given-names>Y.</given-names></string-name>, &amp; <string-name name-style="western"><surname>Loh</surname>, <given-names>W. W.</given-names></string-name></person-group> (<year>2024</year>). <article-title>A structural after measurement approach to structural equation modeling.</article-title> <source>Psychological Methods</source>, <volume>29</volume>(<issue>3</issue>), <fpage>561</fpage>–<lpage>588</lpage>. <pub-id pub-id-type="doi">10.1037/met0000503</pub-id><pub-id pub-id-type="pmid">36355708</pub-id></mixed-citation></ref>
<ref id="r34"><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name name-style="western"><surname>Schwarz</surname>, <given-names>G.</given-names></string-name></person-group> (<year>1978</year>). <article-title>Estimating the dimension of a model.</article-title> <source>Annals of Statistics</source>, <volume>6</volume>(<issue>2</issue>), <fpage>461</fpage>–<lpage>464</lpage>. <pub-id pub-id-type="doi">10.1214/aos/1176344136</pub-id></mixed-citation></ref>
<ref id="r35"><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name name-style="western"><surname>Spiegelhalter</surname>, <given-names>D. J.</given-names></string-name>, <string-name name-style="western"><surname>Best</surname>, <given-names>N. G.</given-names></string-name>, <string-name name-style="western"><surname>Carlin</surname>, <given-names>B. P.</given-names></string-name>, &amp; <string-name name-style="western"><surname>Van Der Linde</surname>, <given-names>A.</given-names></string-name></person-group> (<year>2002</year>). <article-title>Bayesian measures of model complexity and fit.</article-title> <source>Journal of the Royal Statistical Society: Series B (Statistical Methodology)</source>, <volume>64</volume>(<issue>4</issue>), <fpage>583</fpage>–<lpage>639</lpage>. <pub-id pub-id-type="doi">10.1111/1467-9868.00353</pub-id></mixed-citation></ref>
<ref id="r36"><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name name-style="western"><surname>Steinley</surname>, <given-names>D.</given-names></string-name></person-group> (<year>2004</year>). <article-title>Properties of the Hubert-Arable Adjusted Rand Index.</article-title> <source>Psychological Methods</source>, <volume>9</volume>(<issue>3</issue>), <fpage>386</fpage>–<lpage>396</lpage>. <pub-id pub-id-type="doi">10.1037/1082-989X.9.3.386</pub-id><pub-id pub-id-type="pmid">15355155</pub-id></mixed-citation></ref>
<ref id="r37"><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name name-style="western"><surname>Vehtari</surname>, <given-names>A.</given-names></string-name>, <string-name name-style="western"><surname>Gelman</surname>, <given-names>A.</given-names></string-name>, &amp; <string-name name-style="western"><surname>Gabry</surname>, <given-names>J.</given-names></string-name></person-group> (<year>2017</year>). <article-title>Practical Bayesian model evaluation using leave-one-out cross-validation and WAIC.</article-title> <source>Statistics and Computing</source>, <volume>27</volume>, <fpage>1413</fpage>–<lpage>1432</lpage>. <pub-id pub-id-type="doi">10.1007/s11222-016-9696-4</pub-id></mixed-citation></ref>
<ref id="r38"><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name name-style="western"><surname>Vermunt</surname>, <given-names>J. K.</given-names></string-name></person-group> (<year>2025</year>). <article-title>Stepwise estimation of latent variable models: An overview of approaches.</article-title> <source>Statistical Modelling</source>, <volume>25</volume>(<issue>6</issue>), <fpage>530</fpage>–<lpage>551</lpage>. <pub-id pub-id-type="doi">10.1177/1471082X251355693</pub-id></mixed-citation></ref>
<ref id="r39"><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name name-style="western"><surname>Watanabe</surname>, <given-names>S.</given-names></string-name></person-group> (<year>2010</year>). <article-title>Asymptotic equivalence of Bayes cross validation and widely applicable information criterion in singular learning theory.</article-title> <source>Journal of Machine Learning Research</source>, <volume>11</volume>(<issue>116</issue>), <fpage>3571</fpage>–<lpage>3594</lpage>.</mixed-citation></ref>
<ref id="r40"><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name name-style="western"><surname>Wei</surname>, <given-names>X.</given-names></string-name>, <string-name name-style="western"><surname>Huang</surname>, <given-names>J.</given-names></string-name>, <string-name name-style="western"><surname>Zhang</surname>, <given-names>L.</given-names></string-name>, <string-name name-style="western"><surname>Pan</surname>, <given-names>D.</given-names></string-name>, &amp; <string-name name-style="western"><surname>Pan</surname>, <given-names>J.</given-names></string-name></person-group> (<year>2022</year>). <article-title>Evaluation and comparison of SEM, ESEM, and BSEM in estimating structural models with potentially unknown cross-loadings.</article-title> <source>Structural Equation Modeling</source>, <volume>29</volume>(<issue>3</issue>), <fpage>327</fpage>–<lpage>338</lpage>. <pub-id pub-id-type="doi">10.1080/10705511.2021.2006664</pub-id></mixed-citation></ref>
<ref id="r41"><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name name-style="western"><surname>Winter</surname>, <given-names>S. D.</given-names></string-name>, &amp; <string-name name-style="western"><surname>Depaoli</surname>, <given-names>S.</given-names></string-name></person-group> (<year>2020</year>). <article-title>An illustration of Bayesian approximate measurement invariance with longitudinal data and a small sample size.</article-title> <source>International Journal of Behavioral Development</source>, <volume>44</volume>(<issue>4</issue>), <fpage>371</fpage>–<lpage>382</lpage>. <pub-id pub-id-type="doi">10.1177/0165025419880610</pub-id></mixed-citation></ref>
<ref id="r42"><mixed-citation publication-type="web">Zhao, H., Vermunt, J. K., &amp; De Roover, K. (2024). <italic>Novel mixture SEM methods for comparing structural relations among many groups</italic> [OSF project page containing R study codes]. Open Science Framework. <ext-link ext-link-type="uri" xlink:href="https://osf.io/rtp78/">https://osf.io/rtp78/</ext-link></mixed-citation></ref>
<ref id="r43"><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name name-style="western"><surname>Zhao</surname>, <given-names>H.</given-names></string-name>, <string-name name-style="western"><surname>Vermunt</surname>, <given-names>J. K.</given-names></string-name>, &amp; <string-name name-style="western"><surname>De Roover</surname>, <given-names>K.</given-names></string-name></person-group> (<year>2025a</year>). <article-title>Mixture Multilevel SEM vs. Multilevel SEM for comparing structural relations across groups in presence of measurement non-invariance.</article-title> <source>Frontiers in Psychology</source>, <volume>16</volume>, <elocation-id>1463790</elocation-id>. <pub-id pub-id-type="doi">10.3389/fpsyg.2025.1463790</pub-id><pub-id pub-id-type="pmid">40792091</pub-id></mixed-citation></ref>
	<ref id="r44"><mixed-citation publication-type="web">Zhao, H., Vermunt, J. K., &amp; De Roover, K. (2025b). <italic>Supplementary materials to</italic> “Mixture multigroup Bayesian SEM with approximate measurement invariance for comparing structural relations across many groups” [Supplementary materials with ANOVA, bias, and ARI tables; boxplots for ARI and ME, and empirical application of MixMG-BSEM]. PsychOpen GOLD. <pub-id pub-id-type="doi">10.23668/psycharchives.21454</pub-id></mixed-citation></ref>	
</ref-list>
		
	
	<sec sec-type="supplementary-material" id="sp1"><title>Supplementary Materials</title>
		<table-wrap position="anchor">
			<table frame='void' style="background-#f3f3f3">
				<col width="60%" align="left"/>
				<col width="40%" align="left"/>
				<thead>
					<tr>
						<th>Type of supplementary materials</th>
						<th>Availability/Access</th>
					</tr>
				</thead>
				<tbody>
					<tr>
						<th colspan="2">Data</th>						
					</tr>
					<tr>
						<td>No study data is available.</td>
						<td>&mdash;</td>
					</tr>	
					<tr style="grey-border-top-dashed">
						<th colspan="2">Code</th>
					</tr>
					<tr>
						<td>a. R Code - 2MixMG-SEM.</td>
						<td><xref ref-type="bibr" rid="r42">Zhao et al. (2024)</xref></td>
					</tr>
					<tr>
						<td>b. R Code - MixMG-BSEM.</td>
						<td><xref ref-type="bibr" rid="r42">Zhao et al. (2024)</xref></td>
					</tr>
					<tr>
						<td>c. R Code - MixML-SEM.</td>
						<td><xref ref-type="bibr" rid="r42">Zhao et al. (2024)</xref></td>
					</tr>
					<tr>
						<td>d. R Code - Empirical Application.</td>
						<td><xref ref-type="bibr" rid="r42">Zhao et al. (2024)</xref></td>
					</tr>
					<tr style="grey-border-top-dashed">
						<th colspan="2">Material</th>
					</tr>
					<tr>
						<td>a. S1. ANOVA Tables.</td>
						<td><xref ref-type="bibr" rid="r44">Zhao et al. (2025b)</xref></td>
					</tr>
					<tr>
						<td>b. S2. Bias for <inline-formula><mml:math><mml:msub><mml:mrow><mml:mi mathvariant="italic">β</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:math></inline-formula> as an example.</td>
						<td><xref ref-type="bibr" rid="r44">Zhao et al. (2025b)</xref></td>
					</tr>
					<tr>					
						<td>c. S3. Boxplots for ARI and <inline-formula><mml:math><mml:msub><mml:mrow><mml:mi mathvariant="italic">𝑀𝐸</mml:mi></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>𝛽</mml:mi></mml:mrow><mml:mrow><mml:mi>1</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:msub></mml:math></inline-formula>.</td>
						<td><xref ref-type="bibr" rid="r44">Zhao et al. (2025b)</xref></td>
					</tr>
					<tr>
						<td>d. S4. Table for ARI and %CC with Median and MAD.</td>
						<td><xref ref-type="bibr" rid="r44">Zhao et al. (2025b)</xref></td>
					</tr>
					<tr>
						<td>e. S5. Empirical Application of MixMG-BSEM.</td>
						<td><xref ref-type="bibr" rid="r44">Zhao et al. (2025b)</xref></td>
					</tr>
					<tr style="grey-border-top-dashed">
						<th colspan="2">Study/Analysis preregistration</th>
					</tr>	
					<tr>
						<td>The study was not preregistered.</td>
						<td>&mdash;</td>
					</tr>
					<tr style="grey-border-top-dashed">
						<th colspan="2">Other</th>
					</tr>	
					<tr>
						<td>No other material to report.</td>
						<td>&mdash;</td>
					</tr>
				</tbody>
			</table>
		</table-wrap>		
	</sec>
			

<fn-group>
<fn fn-type="conflict"><p>The authors have declared that no competing interests exist.</p></fn>
</fn-group>
<ack>
<p>The authors have no additional (i.e., non-financial) support to report.</p>
</ack>
</back>
</article>
