<?xml version="1.0" encoding="UTF-8"?><rss version="2.0"
	xmlns:content="http://purl.org/rss/1.0/modules/content/"
	xmlns:wfw="http://wellformedweb.org/CommentAPI/"
	xmlns:dc="http://purl.org/dc/elements/1.1/"
	xmlns:atom="http://www.w3.org/2005/Atom"
	xmlns:sy="http://purl.org/rss/1.0/modules/syndication/"
	xmlns:slash="http://purl.org/rss/1.0/modules/slash/"
	>

<channel>
	<title>CenterStat</title>
	<atom:link href="https://centerstat.org/feed/" rel="self" type="application/rss+xml" />
	<link>https://centerstat.org/</link>
	<description>Statistical training in advanced quantitative methods for researchers in the social, health, and behavioral sciences.</description>
	<lastBuildDate>Mon, 30 Mar 2026 18:49:55 +0000</lastBuildDate>
	<language>en-US</language>
	<sy:updatePeriod>
	hourly	</sy:updatePeriod>
	<sy:updateFrequency>
	1	</sy:updateFrequency>
	<generator>https://wordpress.org/?v=6.8.5</generator>

<image>
	<url>https://centerstat.org/wp-content/uploads/2021/01/cropped-centerstat-favicon@4x-32x32.png</url>
	<title>CenterStat</title>
	<link>https://centerstat.org/</link>
	<width>32</width>
	<height>32</height>
</image> 
	<item>
		<title>LatentGOLD demonstrations now included in Mixture Modeling and Latent Class Analysis workshop</title>
		<link>https://centerstat.org/latentgold-demonstrations-now-included-in-mixture-modeling-and-latent-class-analysis-workshop/</link>
		
		<dc:creator><![CDATA[Daniel Bauer, Ph.D.]]></dc:creator>
		<pubDate>Mon, 30 Mar 2026 18:49:55 +0000</pubDate>
				<category><![CDATA[Uncategorized]]></category>
		<category><![CDATA[latent class]]></category>
		<category><![CDATA[latent profile]]></category>
		<category><![CDATA[LatentGOLD]]></category>
		<category><![CDATA[mixture]]></category>
		<guid isPermaLink="false">https://centerstat.org/?p=50554</guid>

					<description><![CDATA[<p>We’re excited to announce that our Mixture Modeling and Latent Class Analysis workshop now includes a full set of demonstrations in the powerful software program&#8230;</p>
<p>The post <a href="https://centerstat.org/latentgold-demonstrations-now-included-in-mixture-modeling-and-latent-class-analysis-workshop/">LatentGOLD demonstrations now included in Mixture Modeling and Latent Class Analysis workshop</a> appeared first on <a href="https://centerstat.org">CenterStat</a>.</p>
]]></description>
										<content:encoded><![CDATA[<p>We’re excited to announce that our <a href="https://centerstat.org/mixture-latent-class/" target="_blank" rel="noopener"><i>Mixture Modeling and Latent Class Analysis</i></a> workshop now includes a full set of demonstrations in the powerful software program <a href="https://www.statisticalinnovations.com/" target="_blank" rel="noopener">LatentGOLD</a>, including data and syntax files, four hours of video recordings, and 120 pages of detailed notes.</p>
<p><img fetchpriority="high" decoding="async" class="alignleft size-full wp-image-50557" src="https://centerstat.org/wp-content/uploads/2026/03/lca_aux4.png" alt="" width="277" height="275" />Why LatentGOLD? First, it is a versatile latent variable modeling software program that can be used to estimate all of the models considered in the class, so why not? Second, not long ago we added an entire chapter to the workshop on approaches for robustly examining relationships between latent classes and external variables (e.g., predictors of class membership or distal outcomes predicted by class membership), and these approaches are exceptionally well implemented in LatentGOLD. This comes as no suprise given the primary developer of LatentGOLD, Jeroen Vermunt, also played a key role in the innovation of these methods. Third, <a href="https://www.statisticalinnovations.com/" target="_blank" rel="noopener">Statistical Innovations Europe</a> now offers a <strong>free</strong> academic license to LatentGOLD and they have partnered with us to provide a 90-day free license to non-academic workshop participants as well.</p>
<p>The LatentGOLD demonstrations complement prior demonstrations in Mplus and R, providing users with a variety of choices when fitting finite mixture models, including applications of latent profile analysis and latent class analysis.</p>
<p>If you previously enrolled in the class, then you have instant access to these materials as a benefit of our <strong><em>evergreen</em></strong> content model. Otherwise, enroll now to take advantage of this added software flexibility!</p>
<p>The post <a href="https://centerstat.org/latentgold-demonstrations-now-included-in-mixture-modeling-and-latent-class-analysis-workshop/">LatentGOLD demonstrations now included in Mixture Modeling and Latent Class Analysis workshop</a> appeared first on <a href="https://centerstat.org">CenterStat</a>.</p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>The Strengths and Limitations of Time-Varying Covariate Growth Models</title>
		<link>https://centerstat.org/the-strengths-and-limitations-of-time-varying-covariate-growth-models/</link>
		
		<dc:creator><![CDATA[Patrick Curran and Dan Bauer]]></dc:creator>
		<pubDate>Sun, 01 Feb 2026 16:17:37 +0000</pubDate>
				<category><![CDATA[Announcement]]></category>
		<category><![CDATA[Help Desk]]></category>
		<category><![CDATA[growth modeling]]></category>
		<category><![CDATA[longitudinal sem]]></category>
		<category><![CDATA[tic]]></category>
		<category><![CDATA[time-invariant covariate]]></category>
		<category><![CDATA[time-varying covariate]]></category>
		<category><![CDATA[tvc]]></category>
		<guid isPermaLink="false">https://centerstat.org/?p=49851</guid>

					<description><![CDATA[<p>Growth models have long been a workhorse of longitudinal data analysis. Whether we are studying reading development across elementary school, depressive symptoms across adulthood, or&#8230;</p>
<p>The post <a href="https://centerstat.org/the-strengths-and-limitations-of-time-varying-covariate-growth-models/">The Strengths and Limitations of Time-Varying Covariate Growth Models</a> appeared first on <a href="https://centerstat.org">CenterStat</a>.</p>
]]></description>
										<content:encoded><![CDATA[
<p>Growth models have long been a workhorse of longitudinal data analysis. Whether we are studying reading development across elementary school, depressive symptoms across adulthood, or political attitudes across election cycles, the core underlying idea is the same: people can change systematically over time, yet they do not all change to the same degree. Despite the tremendous insights offered by growth models, they have a key limitation: they distill variability in <em>within-person</em> (intra-individual) change into coefficients that vary <em>between</em> <em>people</em> (inter-individual differences). That is, some people start higher or lower, and some people increase more or less steeply, but these trajectories are characteristics of the <em>individual</em>: thus, <em>Patrick is six feet tall, weighs 180 pounds, was born in Colorado, and he has an intercept of .5 and a slope of .2</em>. These are all characteristics of <em>Patrick</em>, but none are linked to a specific point in <em>time</em>. We can predict why different people have different intercepts and slopes, but these are all <em>between-person</em> effects. Often, we are also interested in within-person effects: why Patrick was happier today than yesterday, or drank more this year than last.</p>



<p>For decades, we have been able to expand growth models to include time-specific (and even lagged) effects for predictors that themselves change over time, often called time-varying covariates (or TVCs). TVC models allow us to transition to testing <em>within-person</em> hypotheses because the model does not just ask <em>whether</em> people change but whether <em>within-person fluctuations</em> in predictors are related to <em>within-person deviations</em> in the outcome. This is our focus here.</p>



<h3 class="wp-block-heading"><strong>A Brief Review of Growth Modeling</strong></h3>



<p>Growth modeling is a massive topic that cannot be fully covered here. However, we have prior <strong><a href="https://centerstat.org/?s=growth&amp;post_types=post" target="_blank" rel="noreferrer noopener">blog posts</a></strong>, a <strong><a href="https://www.youtube.com/playlist?list=PLQGe6zcSJT0VxMZUN6DBuhIoCRZNoA2Vz" target="_blank" rel="noreferrer noopener">free online lecture series</a></strong> on different approaches to growth modeling, and we offer several full-length workshops<strong> </strong>on growth modeling using <a href="https://centerstat.org/multilevel-models-longitudinal/" target="_blank" rel="noreferrer noopener"><strong>multilevel models</strong></a> (MLMs), <strong><a href="https://centerstat.org/longitudinal-sem/" target="_blank" rel="noreferrer noopener">structural equations models</a></strong> (SEMs), and models for <strong><a href="https://centerstat.org/aild/" target="_blank" rel="noreferrer noopener">intensive longitudinal data</a></strong> (ILD).</p>



<p>Briefly, the core motivating goal of growth modeling, whether estimated using the MLM or the SEM, is to summarize a set of repeated measures with a smoothed underlying trajectory that parsimoniously represents how the outcome changes over time. Individual trajectories can obtain a variety of functional forms (e.g., linear, quadratic, piecewise, exponential, etc.) or even display no systematic change over time (and thus necessitate only an intercept term).</p>



<p>A key outcome of the model is the estimation of the means (sometimes called <em>fixed effects</em>) and the variances (sometimes call <em>random effects</em>) of each defined growth component. For example, say we fit a linear trajectory to six repeated measures of aggressive behavior in children. We could obtain a mean starting point and rate of change reflecting the average trajectory for the full sample, and a variance in starting point and rate of change (and covariance between the two) reflecting child-to-child variability in the trajectory parameters. This model may be of theoretical interest on its own, but we often expand this model to include one or more covariates of theoretical interest. There are two types of covariates that are commonly considered.</p>



<h3 class="wp-block-heading"><strong>Time-invariant vs. Time-varying Covariates</strong></h3>



<p><strong>Time-invariant covariates (TICs)</strong> are assumed to have a constant value across measurement occasions. Examples include sex assigned at birth, baseline SES, treatment condition, or childhood adversity at the start of kindergarten. In growth models, TICs typically predict differences in trajectory parameters and assess questions such as <em>does baseline SES predict the intercept (</em>i.e., <em>starting point)</em> or <em>does treatment condition predict the slope (</em>i.e., <em>rate of change)</em>? An example finding might indicate that there are no treatment group differences in depression at baseline but that, on average, those who received treatment reported a significant decrement in depressive symptomatology over time compared to those who did not. An exemplar path diagram of a latent curve model with two TICs (<em>x</em><sub>1</sub> and <em>x</em><sub>2</sub>) predicting trajectory intercepts and slopes (<em>η</em><sub>1</sub> and <em>η</em><sub>2</sub>) for five repeated measures (<em>y</em><sub>1</sub>&#8211;<em>y</em><sub>5</sub>) is:</p>



<div class="wp-block-uagb-image uagb-block-f0ac84a2 wp-block-uagb-image--layout-default wp-block-uagb-image--effect-static wp-block-uagb-image--align-none"><figure class="wp-block-uagb-image__figure"><img decoding="async" srcset="https://centerstat.org/wp-content/uploads/2026/02/diag1.png ,https://centerstat.org/wp-content/uploads/2026/02/diag1.png 780w, https://centerstat.org/wp-content/uploads/2026/02/diag1.png 360w" sizes="auto, (max-width: 480px) 150px" src="https://centerstat.org/wp-content/uploads/2026/02/diag1.png" alt="" class="uag-image-49861" width="600" height="342" title="diag1" loading="lazy" role="img"/></figure></div>



<p>In contrast, <strong>time-varying covariates (TVCs)</strong> have the potential to differ in value <em>at each measurement occasion</em>. This is a critical distinction from TICs. Whereas baseline SES is by definition invariant to the passage of time, other constructs, such as daily stress, nightly sleep, current caregiver status, or weekly anxiety, change from one time point to the next. Whereas TICs predict the growth factors themselves (e.g., we regress the intercept and slope trajectory components on treatment group membership), TVCs directly predict the time-specific repeated assessments <em>above-and-beyond</em> the underlying growth process. Although a seemingly modest change to the structure of the model, this actually has profound implications for the types of questions we can ask, particularly those focused on within-person process. An exemplar path diagram of a latent curve model with a TVC (<em>z</em><sub>1</sub>&#8211;<em>z</em><sub>5</sub> measured at time 1 to 5) is:</p>



<div class="wp-block-uagb-image uagb-block-7154373e wp-block-uagb-image--layout-default wp-block-uagb-image--effect-static wp-block-uagb-image--align-none"><figure class="wp-block-uagb-image__figure"><img decoding="async" srcset="https://centerstat.org/wp-content/uploads/2026/02/diag2.png ,https://centerstat.org/wp-content/uploads/2026/02/diag2.png 780w, https://centerstat.org/wp-content/uploads/2026/02/diag2.png 360w" sizes="auto, (max-width: 480px) 150px" src="https://centerstat.org/wp-content/uploads/2026/02/diag2.png" alt="" class="uag-image-49862" width="525" height="404" title="diag2" loading="lazy" role="img"/></figure></div>



<p>Note that TVCs can have both within- and between-person effects.&nbsp; For example, suppose we wish to evaluate the relation between substance use and anxiety over a set of repeated measures. &nbsp;The between-person effect would reflect average differences between people: <em>Do people who report higher average levels of anxiety over time also report higher average levels of substance use over time</em>? This contrasts from the within-person effect: <em>When a person is more anxious <u>than usual</u> at a given time point do they tend to engage in more substance use <u>than they typically do</u></em>? Often with TVCs we are particularly interested in modeling these within-person processes, which can be extended to examine lead-lag relations, and dynamic systems in ways not possible with a standard TIC model.</p>



<p>In the spirit that you don&#8217;t get something for nothing, inclusion of TVCs introduces natural complexities that must be addressed, particularly in the management of the data, specification of the model, and substantive interpretation of the findings. However, these issues are all well understood and one needs to simply turn to existing resources that illustrate best practices.</p>



<h3 class="wp-block-heading"><strong>Incorporating TVCs into Growth Models</strong></h3>



<p>One point of common confusion is how within- and between-person effects of TVCs are estimated and interpreted within the SEM versus MLM frameworks. Within the SEM growth model, the TVCs are incorporated in their raw metric as exogenous predictors of the repeated measures. Because the TVCs freely covary with the latent growth factors (or at least should; be sure your software package is estimating these relations), the resulting regression parameters represent pure estimates of within-person effects. The standard SEM specification does not provide pure estimates of between-person effects of TVCs; however, these can be obtained by expanding the model in particular ways (see Curran et al., 2012, for details).</p>



<p>In contrast, if the TVCs are incorporated into the MLM in their raw metric, the resulting regression coefficients inextricably combine the within-person and between-person effects (see Raudenbush &amp; Bryk, 2002, for details). A well-known solution to this problem is to person-mean center, that is subtract the person mean of the TVC from each observed value of the TVC, and then use the person-mean-centered TVC as the predictor at Level 1; this captures the same pure within-person effect estimate as the SEM. However, the person mean of the TVC itself can also be incorporated as an additional predictor at Level 2 of the model to capture the between-effect.</p>



<p>Thus, under a broad set of conditions, the MLM and SEM will provide precisely the same (or nearly the same) estimates of the within and between effects of TVCs, even though this is accomplished in quite different ways.</p>



<h3 class="wp-block-heading"><strong>Limitations</strong></h3>



<p>Both the SEM and MLM approaches to modeling TVCs include a number of assumptions, but two are of most interest here. First, using either of the methods described above, we make a fundamental assumption that although the distal outcome may be growing systematically over time, the TVC most decidedly is <em>not</em>. Think about this logically: we deviate the person mean from each measured TVC and the person mean is by definition constant over time. In other words, there is no growth trajectory for the TVC. Significant estimation and interpretation problems can be encountered if the TVC <em>itself</em> is growing over time and this is not adequately represented in the model (see Curran &amp; Bauer, 2011, for detailed examples, and Wang &amp; Maxwell, 2015, for exceptions). To be clear, methods exist for handling such conditions, but additional work is necessary to represent systematic co-occurring individual change in the model (Curran et al., 2014).</p>



<p>The second assumption relates to the strength of relation between the TVC and the outcome over time. Consider three possible scenarios: (1) the magnitude of the relation is equal at all time points; (2) the magnitude of the relation systematically increases or decreases with the passage of time; or (3) the magnitude of the relation obtains a unique value at every single time point. All three of these can be incorporated in both the SEM and the MLM, and formal tests are available to determine which condition best represents the sample data at hand. However, although these methods offer powerful tests of the nature of the relation between the TVC and the outcome, there is a well-developed yet little used method that provides more flexible insight into these over-time relations, and this is called the time-varying effects model, or TVEM.</p>



<h3 class="wp-block-heading"><strong>Time-Varying Effects Models</strong></h3>



<p>Time-varying effects models offer a semi-parametric approach to the more traditional TVC model within the MLM framework. Whereas the MLM provides formal tests of structured relations between the TVC and the outcome (e.g., a bilinear “product” interaction between the TVC and time), the TVEM uses a spline method of estimation to approximate complex nonlinear relations that might wax and wane in magnitude over time. Confidence regions can then be plotted to show &#8220;sensitive periods&#8221; during which a TVC is significantly related to the outcome versus periods when it is not. These relations are primarily graphical in nature and lovely plots can be created that demonstrate the potentially complex nature of the relation between the TVC and the outcome as a function of time. The TVEM offers many advantages when trying to understand complex relations over time, yet these methods remain quite underutilized in practice. See Lanza and Linden-Carmichael (2021) for a thorough introduction to the estimation and interpretation of TVEMs within the social sciences.</p>



<h3 class="wp-block-heading"><strong>Conclusion</strong></h3>



<p>In sum, time-varying covariate growth models allow for the introduction of powerful tests of within-person dynamics over time that are not present in a more standard TIC. These methods are well developed and widely used and can easily be incorporated in your own data applications. A variety of resources exist, and we cite several of these below. CenterStat also offers full workshops that provide detailed instruction on TVC models, including <strong><em><a href="https://centerstat.org/multilevel-models-longitudinal/" target="_blank" rel="noreferrer noopener">Multilevel Models for Longitudinal Data</a></em></strong>, <strong><em><a href="https://centerstat.org/longitudinal-sem/" target="_blank" rel="noreferrer noopener">Longitudinal Structural Equation Modeling</a></em></strong>, and <strong><em><a href="https://centerstat.org/aild/" target="_blank" rel="noreferrer noopener">Analysis of Intensive Longitudinal Data</a></em></strong><em>. </em>Including TVCs in your models requires care but can also yield important new insights.</p>



<h3 class="wp-block-heading"><strong>Suggested Readings</strong></h3>



<p>Curran, P. J., &amp; Bauer, D. J. (2011). The disaggregation of within-person and between-person effects in longitudinal models of change.&nbsp;<em>Annual Review of Psychology</em>,&nbsp;<em>62</em>, 583-619.</p>



<p>Curran, P. J., Howard, A. L., Bainter, S. A., Lane, S. T., &amp; McGinley, J. S. (2014). The separation of between-person and within-person components of individual change over time: a latent curve model with structured residuals.&nbsp;<em>Journal of consulting and clinical psychology</em>,&nbsp;<em>82</em>(5), 879.</p>



<p>Curran, P.J., Lee, T.H., Howard, A.H., Lane, S.T., &amp; MacCallum, R.C. (2012). Disaggregating within-person and between-person effects in multilevel and structural equation growth models. In: Hancock, G., editor. Advances in longitudinal methods in the social and behavioral sciences. Charlotte, NC: Information Age; p. 217-253.</p>



<p>Hamaker, E. L., Kuiper, R. M., &amp; Grasman, R. P. (2015). A critique of the cross-lagged panel model.&nbsp;<em>Psychological Methods</em>,&nbsp;<em>20</em>(1), 102.</p>



<p>Lanza, S. T., &amp; Linden-Carmichael, A. N. (2021).&nbsp;<em>Time-varying effect modeling for the behavioral, social, and health sciences</em>. Springer.</p>



<p>Raudenbush, S.W. &amp; Bryk, A.S. (2002). Hierarchical linear models: Applications and data analysis methods.&nbsp;<em>Advanced Quantitative Techniques in the Social Sciences Series/SAGE</em>.</p>



<p>Shiyko, M.P., Burkhalter, J., Li, R., &amp; Park, B. J. (2014). Modeling nonlinear time-dependent treatment effects: an application of the generalized time-varying effect model (TVEM).&nbsp;<em>Journal of Consulting and Clinical Psychology</em>,&nbsp;<em>82</em>, 760.</p>



<p>Wang L.P. &amp; Maxwell S.E. (2015). On disaggregating between-person and within-person effects with longitudinal data using multilevel models. <em>Psychological Methods, 20</em>, 63-83.</p>



<p>&nbsp;</p>
<p>The post <a href="https://centerstat.org/the-strengths-and-limitations-of-time-varying-covariate-growth-models/">The Strengths and Limitations of Time-Varying Covariate Growth Models</a> appeared first on <a href="https://centerstat.org">CenterStat</a>.</p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>Understanding the Bootstrap</title>
		<link>https://centerstat.org/understanding-the-bootstrap/</link>
		
		<dc:creator><![CDATA[Patrick Curran and Dan Bauer]]></dc:creator>
		<pubDate>Wed, 17 Sep 2025 15:24:37 +0000</pubDate>
				<category><![CDATA[Announcement]]></category>
		<category><![CDATA[Help Desk]]></category>
		<guid isPermaLink="false">https://centerstat.org/?p=48827</guid>

					<description><![CDATA[<p>In modern research, one of the most fundamental challenges is uncertainty. Whenever we collect data, whether from surveys, experiments, or observational studies, we want to&#8230;</p>
<p>The post <a href="https://centerstat.org/understanding-the-bootstrap/">Understanding the Bootstrap</a> appeared first on <a href="https://centerstat.org">CenterStat</a>.</p>
]]></description>
										<content:encoded><![CDATA[
<p>In modern research, one of the most fundamental challenges is uncertainty. Whenever we collect data, whether from surveys, experiments, or observational studies, we want to make claims not only about the specific sample we observe but about the broader population it represents. Doing this requires tools for statistical inference, and central to inference is the concept of a <em>sampling distribution</em>. Traditionally, researchers have relied on parametric approaches, but these often invoke strong assumptions that may not hold in practice. However, in the last half-century the <em>bootstrap</em> has emerged as one of the most influential methods for estimating sampling variability in a way that requires far fewer assumptions and is now widely accessible given advances in high speed computing. Although there are many “flavors” of bootstrapping, our focus here is on the non-parametric bootstrap which involves resampling raw data from the original observed sample values.</p>



<h3 class="wp-block-heading">Parametric Inference and the Sampling Distribution</h3>



<p>In classical statistics, inference depends on specifying a model for the data-generating process and the idea of repeated sampling. For instance, suppose we take a sample of size <em>N</em> and calculate the sample mean. Ultimately, we’d like to make a conclusion about the mean in the population from which we drew our sample. However, we need to account for sampling error. That is, had we drawn a different sample of the same size, we would have obtained a different sample mean. In fact, there are infinitely many different samples of size <em>N</em> that we might have drawn from our population, each of which would yield a somewhat different mean, simply depending on who happened to get into the sample. The collection of means across all of these hypothetical samples has its own distribution, known as a <em>sampling distribution</em>. The sampling distribution reflects our uncertainty in the estimation of the population mean (i.e., variation due to sampling error) and it underpins confidence intervals, hypothesis tests, and other inferential procedures.</p>



<p>For instance, under the assumption that the we draw independent and identically distributed observations from a normal distribution (or by appealing to the Central Limit Theorem at large sample sizes), the sample mean follows a normal distribution with mean equal to the population mean and standard error equal to the standard deviation divided by the square root of <em>N</em>. If the standard deviation for the population is known, then we can simply divide our sample mean by the standard error and reference this to the standard normal distribution (<em>z</em>-distribution) to make inferences. More typically, however, we aren’t privy to this knowledge and must plug in our sample standard deviation to get the standard error. The sampling distribution of the mean will then deviate from the normal curve due to this added uncertainty in small samples. Thanks to our favorite Guinness brewer William Gossett (who published under &#8220;Student&#8221;), we know that in such cases, the sample mean divided by the standard error follows a <em>t</em>-distribution, a bell-shaped distribution whose tail thickness is determined by the degrees-of-freedom.</p>



<p>So far, we have focused on the sample mean, but we can imagine a sampling distribution for any parameter we wish to estimate, whether it be a regression coefficient, variance, factor loading, or any other value of interest. The challenge, however, is that in real research we often find ourselves making uncertain parametric assumptions in order to obtain a known sampling distribution. For example, when using a <em>t</em>-distribution to make inferences about a regression coefficient, we assume a sufficiently large sample size and that the errors are normally distributed, independent, and homoscedastic, conditions that may or may not hold in practice. When assumptions are met, such parametric procedures work exceedingly well; when not met, the resulting inferences can be both biased and misleading, sometimes markedly so.</p>



<h3 class="wp-block-heading">What Is the Bootstrap?</h3>



<p>The non-parametric bootstrap, first formally proposed by Bradley Efron in 1979, is a computational technique for <em>empirically approximating</em> the sampling distribution without the requirement of strong parametric assumptions. Instead of relying on mathematical formulas, the bootstrap uses the observed data itself as a stand-in for the population. (Thus, the term &#8220;bootstrap&#8221; which is drawn from the phrase <em>draw yourself up by your own bootstraps</em> meaning you take personal responsibility and use the resources you have available to you). The basic procedure is conceptually simple. First, you draw your sample of size <em>N</em> from the population in the usual way. Next, you draw a “bootstrap sample” also of size <em>N</em> from the original data <em>with replacement</em>. Then you compute and retain your statistic of interest on the bootstrap sample in whatever way you please (e.g., mean, regression coefficient, mediated effect). Finally, you repeat this process many times (often with 1000 bootstrap samples or more) to create an empirical distribution for the statistic from which to make inferences back to the population.</p>



<p>The critical step to understand is that we are randomly drawing a bootstrap sample from our original sample data with replacement. Say we have a sample of <em>N</em>=100 observations; we would draw say 1000 bootstrap samples of size 100 where in each one a given observation may appear repeatedly or not at all (thus the sampling &#8220;with replacement&#8221;). This empirical distribution of the estimate (whatever that might be) under very general conditions approximates the sampling distribution. As such, we can use this to compute standard errors, confidence intervals, and bias estimates in similar ways to that of the parametric sampling distribution but with far fewer assumptions about the population. For instance, the bootstrapped standard error is simply the standard deviation of the bootstrapped sample estimates. And a bootstrapped confidence interval can be computed by simply locating the 2.5<sup>th</sup> and 97.5<sup>th</sup> percentiles of the bootstrapped sample estimates, values that may or may not be symmetric around the estimate (in contrast to the symmetry assumed by parametric <em>z</em>&#8211; or <em>t</em>-type confidence intervals). The beauty of the bootstrap lies in how it transforms a theoretical problem (deriving the sampling distribution) into a computational one.</p>



<h3 class="wp-block-heading">Typical Applications</h3>



<p>The bootstrap has found applications across nearly every domain of research. A classic example in the social sciences relates to the testing of indirect effects in mediation models, path analysis, and structural equation modeling. Such an effect arises within a causal chain within which one variable affects another which in turn affects a third (with more complex chains also being possible). Each effect within the causal chain is represented by its own regression coefficient. Under standard assumptions, and in large samples, each regression coefficient estimate will have a normal sampling distribution, allowing for the usual inferences. However, we don’t want to test each link in the chain individually; we want to test the chain as a whole. That is, we want to test the indirect effect of the initial predictor on the final outcome as transmitted through the intervening variables (or mediators). The sample estimate of an indirect effect is obtained by computing the product of the regression coefficients involved in the chain. Easy enough. To test the indirect effect, however, we need to know its sampling distribution, and that’s where things get tricky. Each link in the chain has a normal sampling distribution, but a product of normal variates is generally not itself normally distributed. Using a normal sampling distribution as an approximation (the delta-method or “Sobel method” for testing indirect effects) is convenient but often leads to biased inferences. This is a well-known problem that has sparked a variety of solutions, one of which is to derive the correct parametric distribution for the indirect effect (known as the “distribution of the product” method). More commonly, however, investigators have turned to the non-parametric bootstrap to obtain empirically-based inferential tests that do not rely on parametric assumptions at all. Indeed, bootstrapping is now the gold standard for testing mediated effects in practice.</p>



<p>Regardless of whether one is evaluating an indirect effect, a variance estimate, or any sample estimate of interest, there are many potential uses of the bootstrap results. For example, we can estimate standard errors, particularly when no simple formula exists (e.g., in complex nonlinear models). Similarly, we can compute confidence intervals using several different bootstrap methods (percentile, bias-corrected, accelerated) that provide intervals with better coverage properties than parametric ones in certain settings. Further, in machine learning and predictive modeling, bootstrap samples can be used for model validation and cross-validation and to estimate prediction error. Finally, in fields where data collection is difficult or expensive, such as clinical trials, educational experiments, or niche social science surveys, the bootstrap offers a way to make inference with small samples and limited data. These are just a few examples of how the bootstrap can be used in practice, and many additional options are available.</p>



<h3 class="wp-block-heading">Advantages of the Bootstrap</h3>



<p>There are many advantages to the bootstrap. Unlike traditional parametric methods, the non-parametric bootstrap does not require specifying a functional form for the population distribution. This makes it attractive when normality or homoscedasticity is questionable. The method also works for a wide range of statistics including means, medians, regression coefficients, correlation coefficients, or even more complex estimands like Gini coefficients. At its core, the bootstrap is easy to explain and implement. With modern software (R, Stata, SAS, Python), the procedure often requires just a few lines of code. The bootstrap can be extended for clustered data, time series, or hierarchical designs (e.g., students nested within classrooms), making it useful in applied social science and education research. As a general method, the bootstrap is remarkably flexible and can be applied in many interesting and challenging research scenarios.</p>



<h3 class="wp-block-heading">Disadvantages and Limitations</h3>



<p>As with any procedure, there are also disadvantages that must be considered. Because the bootstrap treats the sample as a proxy for the population, any biases in the sample will propagate through the bootstrap distribution. In small or unrepresentative samples, the bootstrap may give misleading results. Although less of an issue today, the bootstrap can be computationally intensive, especially for large datasets or complex models. Thousands of resamples are often needed for stable estimates. The bootstrap may also perform poorly for statistics that depend heavily on the tails of the distribution (e.g., extreme quantiles, maximum values), because the resampled datasets cannot create values outside the observed range. In clustered or dependent data structures, naïve bootstrapping can underestimate variability unless modified (e.g., block bootstrap, cluster bootstrap). This is particularly relevant in education research, where students are not independent observations, or repeated measures applications, where observations are correlated over time within persons. Care must be taken when evaluating the potential use of the bootstrap procedure in practice given the associated limitations.</p>



<h3 class="wp-block-heading">Conclusion</h3>



<p>The bootstrap represents one of the great innovations in modern statistics: a method that converts inference from an algebraic to a computational problem. By resampling from the observed data, researchers can approximate the sampling distribution of almost any statistic, gaining access to standard errors, confidence intervals, and bias estimates without heavy reliance on parametric formulas. Its strengths (flexibility, fewer assumptions, and ease of implementation) make it a powerful tool, especially in social sciences and education where data are often messy, distributions non-normal, and sample sizes modest. Yet, it is not a panacea: bootstrap inference depends on sample representativeness, can be computationally costly, and struggles with extreme statistics or dependent data if applied naïvely. For applied researchers, the bootstrap is best viewed as one tool in the inferential toolbox. When combined with sound research design and thoughtful modeling, it provides a robust way to grapple with uncertainty and to extract credible insights from limited data.</p>



<h3 class="wp-block-heading">Suggested Readings</h3>



<p>Alfons, A., Ateş, N. Y., &amp; Groenen, P. J. (2022). A robust bootstrap test for mediation analysis. <em>Organizational Research Methods, 25</em>, 591-617.</p>



<p>Efron, B. (1979). Bootstrap Methods: Another look at the jackknife. <em>The Annals of Statistics, 7, </em>1-26.</p>



<p>Efron, B. (2000). The bootstrap and modern statistics. J<em>ournal of the American Statistical Association</em>, 95, 1293-1296.</p>



<p>Efron, B., &amp; Tibshirani, R. (1986). Bootstrap methods for standard errors, confidence intervals, and other measures of statistical accuracy. <em>Statistical Science</em>, 54-75.</p>



<p>McLachlan, G.J. (1987). On bootstrapping the likelihood ratio test statistic for the number of components in a normal mixture. <em>Journal of the Royal Statistical Society, Series C, 36</em>, 318-324.</p>



<p>Preacher, K. J., &amp; Hayes, A. F. (2008). Asymptotic and resampling strategies for assessing and comparing indirect effects in multiple mediator models. <em>Behavior Research Methods, 40</em>, 879-891.</p>



<p>Stine, R. (1989). An introduction to bootstrap methods: Examples and ideas. <em>Sociological Methods &amp; Research, 18</em>, 243-291.</p>



<p>Tibshirani, R. J., &amp; Efron, B. (1993). An introduction to the bootstrap. <em>Monographs on Statistics and Applied Probability, 57</em>, 1-436.</p>



<p>&nbsp;</p>
<p>The post <a href="https://centerstat.org/understanding-the-bootstrap/">Understanding the Bootstrap</a> appeared first on <a href="https://centerstat.org">CenterStat</a>.</p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>Discover the Power of Qualitative Research</title>
		<link>https://centerstat.org/discover-the-power-of-qualitative-research/</link>
		
		<dc:creator><![CDATA[Patrick Curran and Dan Bauer]]></dc:creator>
		<pubDate>Sat, 16 Aug 2025 18:27:00 +0000</pubDate>
				<category><![CDATA[Announcement]]></category>
		<category><![CDATA[Help Desk]]></category>
		<guid isPermaLink="false">https://centerstat.org/?p=48270</guid>

					<description><![CDATA[<p>CenterStat Launches Four New Applied Workshops In a world overflowing with data, it’s easy to assume that numbers tell the whole story. But while quantitative&#8230;</p>
<p>The post <a href="https://centerstat.org/discover-the-power-of-qualitative-research/">Discover the Power of Qualitative Research</a> appeared first on <a href="https://centerstat.org">CenterStat</a>.</p>
]]></description>
										<content:encoded><![CDATA[
<h2 class="wp-block-heading"><strong>CenterStat Launches Four New Applied Workshops</strong></h2>



<p>In a world overflowing with data, it’s easy to assume that numbers tell the whole story. But while quantitative data can show <em>what</em> is happening, it rarely explains <em>why</em>. That’s where qualitative research comes in—offering the rich, nuanced insights needed to truly understand human behavior, decision-making, and the social and cultural contexts that shape them.</p>



<p>Qualitative research goes beyond surface-level observations to uncover motivations, beliefs, and lived experiences. Through interviews, focus groups, observations, and textual analysis, it gives researchers the tools to see the world from the perspective of the people living in it. These methods have the power to transform the way we design programs, implement policies, and improve services across fields like healthcare, education, business, and community development.</p>



<p>And now, <strong>CenterStat is making it easier than ever to master these skills</strong> with the launch of the <strong>Applied Qualitative Research (AQR)</strong> workshop series—a comprehensive, four-part training program designed for both new and experienced researchers.</p>



<h2 class="wp-block-heading"><strong>Why Qualitative Research Matters More Than Ever</strong></h2>



<p>In applied research and the social sciences, understanding the “human side” of the data is essential. Qualitative research provides:</p>



<ul class="wp-block-list">
<li><strong>Deeper motivations</strong> – Find out the reasons behind actions, attitudes, and behaviors.</li>



<li><strong>Context-rich insights</strong> – Understand how cultural norms, social structures, and personal experiences influence outcomes.</li>



<li><strong>Flexibility</strong> – Adapt your research as new themes and discoveries emerge.</li>



<li><strong>Participant-centered perspectives</strong> – Amplify the voices of underrepresented or marginalized communities.</li>



<li><strong>Stronger quantitative studies</strong> – Use qualitative methods to clarify constructs, refine survey questions, and build better measurement tools.</li>
</ul>



<p>For example, in healthcare, interviews with patients can reveal emotional responses to treatment or cultural barriers to care—details that can lead to more effective and equitable interventions. In organizational research, focus groups might uncover hidden resistance to change or shed light on team dynamics that numbers alone could never explain.</p>



<h2 class="wp-block-heading"><strong>A Flexible, Practical Approach to Learning</strong></h2>



<p>CenterStat’s <strong>Applied Qualitative Research</strong> series is designed with real-world application in mind. Whether you work in academia, policy, healthcare, nonprofit organizations, or private industry, you’ll walk away with skills you can put to work immediately.</p>



<p>Each workshop combines clear, research-based instruction with practical exercises, templates, and tools. You can take the full series for a complete learning experience or choose the individual sessions that best fit your needs.</p>



<h2 class="wp-block-heading"><strong>The Four Workshops in the AQR Series</strong></h2>



<h3 class="wp-block-heading"><strong>1. Applied Qualitative Research: Foundations </strong><strong><em>(Free Introductory Session)</em></strong></h3>



<p>Your starting point for mastering qualitative research. This session introduces you to the core principles of qualitative inquiry, including how it fits within a pragmatist framework. You’ll learn to identify the right research questions, select appropriate study populations, and define units of observation. The highlight is <strong>The Nine Elements of a Good Research Question</strong>, a practical tool that ensures your research starts on the right track.</p>



<h3 class="wp-block-heading"><strong>2. In-Depth Interviews (IDIs)</strong></h3>



<p>Learn to conduct powerful, one-on-one conversations that uncover authentic insights. This workshop covers:</p>



<ul class="wp-block-list">
<li>When and why to use IDIs</li>



<li>Sampling and recruitment strategies</li>



<li>Interview guide development</li>



<li>The art of inductive probing to elicit deeper responses</li>



<li>Practical tips for in-person and remote interviews</li>
</ul>



<p>Participants will also observe and critique a demonstration interview, and receive ready-to-use tools such as transcription protocols and data management templates.</p>



<h3 class="wp-block-heading"><strong>3. Focus Groups (FGs)</strong></h3>



<p>Harness the dynamic energy of group conversation to explore shared and differing perspectives. You’ll learn:</p>



<ul class="wp-block-list">
<li>How to recruit participants and determine sample sizes</li>



<li>Best practices for developing an FG discussion guide</li>



<li>Strategies for moderating discussions and managing group dynamics</li>



<li>How to work effectively with co-facilitators and assistants</li>



<li>Special considerations for running digital focus groups</li>
</ul>



<p>The session includes a demonstration FG and practical templates to simplify documentation and transcription.</p>



<h3 class="wp-block-heading"><strong>4. Thematic Analysis</strong></h3>



<p>Transform your raw qualitative data into clear, actionable findings. This hands-on workshop takes you step-by-step through:</p>



<ul class="wp-block-list">
<li>Coding and identifying themes</li>



<li>Building a codebook and linking themes into a conceptual model</li>



<li>Using qualitative data analysis (QDA) software</li>



<li>Writing results for academic and applied audiences</li>
</ul>



<p>The focus is on producing research that is both rigorous and ready to make an impact.</p>



<h2 class="wp-block-heading"><strong>Why Choose CenterStat for Qualitative Training?</strong></h2>



<ul class="wp-block-list">
<li><strong>Expert Instructors</strong> – Learn from leaders in statistical and qualitative research education with years of hands-on experience.</li>



<li><strong>Practical Tools</strong> – Every session includes ready-to-use templates, protocols, and guides.</li>



<li><strong>Flexible Learning</strong> – Take the workshops in sequence or choose the ones that meet your current needs.</li>



<li><strong>Designed for All Levels</strong> – From beginners to seasoned researchers, every participant gains actionable skills.</li>



<li><strong>Real-World Relevance</strong> – Training is tailored for applied research, not just theory.</li>
</ul>



<h2 class="wp-block-heading"><strong>Make Your Research Count</strong></h2>



<p>In an era where data is everywhere, the ability to truly understand people—beyond the numbers—is a competitive advantage. The skills you develop through the <strong>Applied Qualitative Research</strong> series can help you design more effective studies, communicate your findings with greater impact, and ultimately create change that matters.</p>



<p>Whether you are launching your first qualitative study or seeking to expand your analytical toolkit, this series will give you the confidence, clarity, and capability to succeed.</p>



<p><strong>Ready to get started?</strong> See upcoming sessions and register today at<a href="https://centerstat.org"> CenterStat.org</a>.</p>



<p>&nbsp;</p>
<p>The post <a href="https://centerstat.org/discover-the-power-of-qualitative-research/">Discover the Power of Qualitative Research</a> appeared first on <a href="https://centerstat.org">CenterStat</a>.</p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>Why CenterStat is the Best Choice for Online Statistics Training</title>
		<link>https://centerstat.org/why-centerstat-is-the-best-choice-for-online-statistics-training/</link>
		
		<dc:creator><![CDATA[Patrick Curran and Dan Bauer]]></dc:creator>
		<pubDate>Mon, 16 Jun 2025 17:26:16 +0000</pubDate>
				<category><![CDATA[Announcement]]></category>
		<category><![CDATA[Help Desk]]></category>
		<guid isPermaLink="false">https://centerstat.org/?p=46475</guid>

					<description><![CDATA[<p>In today’s data-driven world, statistics isn’t just a tool: it’s a necessity. Whether you’re a researcher, graduate student, data analyst, or academic professional, high-quality statistical&#8230;</p>
<p>The post <a href="https://centerstat.org/why-centerstat-is-the-best-choice-for-online-statistics-training/">Why CenterStat is the Best Choice for Online Statistics Training</a> appeared first on <a href="https://centerstat.org">CenterStat</a>.</p>
]]></description>
										<content:encoded><![CDATA[
<p>In today’s data-driven world, statistics isn’t just a tool: it’s a necessity. Whether you’re a researcher, graduate student, data analyst, or academic professional, high-quality statistical training can dramatically advance your skills, open doors to new opportunities, and make stronger scientific contributions to society. But with countless online providers offering statistics workshops and courses, how can you decide which one is truly worth your time and investment?</p>



<p>At <strong>CenterStat</strong>, we believe you shouldn’t have to choose between quality and affordability. That’s why we’ve built an online education platform that delivers <strong>unmatched excellence, clarity, and value</strong>. When compared with other providers, CenterStat doesn’t just compete: we lead.</p>



<p>Here’s why CenterStat is the gold standard for online statistics training, and why thousands of learners around the world choose us as their trusted resource.</p>



<hr class="wp-block-separator has-alpha-channel-opacity"/>



<p><strong>1. The Most Affordable Cost for the Highest Quality Training</strong></p>



<p>Let’s face it: many online courses in statistics and data science come with a hefty price tag, often far exceeding their value. At CenterStat, our mission is different. We believe that <strong>world-class training should be accessible to everyone</strong>, not just those with large research budgets or institutional funding.</p>



<p>That’s why we’ve made our tuition structure among the most affordable in the industry—without compromising on quality. We deliver the depth, rigor, and expertise of a university-level course at a fraction of the cost. It’s not just a better price—it’s <strong>better value</strong>.</p>



<hr class="wp-block-separator has-alpha-channel-opacity"/>



<p><strong>2. Instructors Who Are Leaders in Their Fields</strong></p>



<p>Our instructors aren’t just skilled educators—they’re <strong>internationally recognized experts</strong> who have made <strong>novel contributions to quantitative methods</strong> in their respective fields. Every member of our instructional team is deeply experienced, actively publishing, and helping shape the landscape of modern statistical research.</p>



<p>Moreover, many of our faculty have won <strong>university-wide and national awards</strong> for both teaching and research. They bring that same passion for clarity, depth, and practical application to every workshop we offer.</p>



<hr class="wp-block-separator has-alpha-channel-opacity"/>



<p><strong>3. High-Quality Materials with a Consistent Structure</strong></p>



<p>One common complaint about online courses is inconsistency—each course is built differently, leaving learners confused and frustrated. At CenterStat, we take a different approach. Every course is developed within a <strong>carefully designed, consistent framework</strong>.</p>



<p>Each class includes comprehensive <strong>PDF lecture notes</strong>, step-by-step <strong>computer demonstration notes</strong>, fully executable <strong>computer code</strong> and real-world <strong>sample datasets</strong>. This level of organization allows you to focus on learning—not on figuring out how the materials fit together. Participants are confident that they will get the same level of quality and comprehensiveness no matter which classes in which they choose to enroll.</p>



<hr class="wp-block-separator has-alpha-channel-opacity"/>



<p><strong>4. Self-Paced, Individually Pre-Recorded Video Instruction</strong></p>



<p>Unlike many competitors who simply upload recordings of live classes, CenterStat creates <strong>custom, pre-recorded videos</strong> specifically designed for self-paced learning. This means there are no audio glitches or disorganized transitions, there are focused explanations tailored to asynchronous learners, and there are clear visual demonstrations, edited for maximum clarity. The result is a <strong>far more effective and engaging learning experience</strong> than when a live offering is simply recorded and posted for later viewing.</p>



<hr class="wp-block-separator has-alpha-channel-opacity"/>



<p><strong>5. Real-Data Demonstrations Across All Major Software</strong></p>



<p>Theory is important, but application is critical. At CenterStat, all of our courses include <strong>hands-on demonstrations using real data</strong>, not contrived classroom examples. Even better, we support instruction across <strong>all major statistical platforms</strong> including R, SAS, Stata, SPSS, Mplus, and Python. No matter what tools you use (or new tools you want to learn), our courses meet you where you are and guide you in developing skills that are <strong>immediately transferable</strong> to your own research or professional projects.</p>



<hr class="wp-block-separator has-alpha-channel-opacity"/>



<p><strong>6. Lifetime Access, No Expirations: Ever. You heard us right: Ever.&nbsp;</strong></p>



<p>When you register for a CenterStat course, <strong>you own it for life</strong>. All videos, PDFs, code, and datasets remain available indefinitely—so you can revisit the material anytime, at your convenience.</p>



<p>Whether you want to refresh a method in a year or re-watch a demonstration before applying it to new data, CenterStat is always there to support you.</p>



<hr class="wp-block-separator has-alpha-channel-opacity"/>



<p><strong>7. Evergreen Content with Free Updates</strong></p>



<p>Most online training platforms treat courses as fixed products: once you buy them, any improvements or expansions are locked behind new fees. Not at CenterStat.</p>



<p>We treat all our major content as <strong>evergreen</strong>. That means if a course is updated, <strong>you get the new content automatically</strong>, at no additional cost—regardless of when you registered. You’ll always have access to the most current, relevant training in the field.</p>



<hr class="wp-block-separator has-alpha-channel-opacity"/>



<p><strong>8. A Commitment to Free, Public Educational Resources</strong></p>



<p>At CenterStat, we’re committed not just to our students, but to the <strong>broader public good</strong>. That’s why we’ve created and maintain a rich library of <strong>free instructional materials </strong>including extensive tutorial videos, lecture notes, coding tutorials, sample code and data, reference guides, and more.</p>



<p>We believe deeply in the <strong>broadest possible dissemination of quantitative education</strong>, and we make meaningful, high-quality resources available to everyone, regardless of financial status.</p>



<hr class="wp-block-separator has-alpha-channel-opacity"/>



<p><strong>9. A Fully Integrated Curriculum</strong></p>



<p>Most online training sites offer a scattered menu of unrelated courses created by different instructors with no overarching plan or coordination. CenterStat offers something better: a <strong>carefully integrated curriculum</strong>.</p>



<p>Every course is designed to <strong>fit within a coherent learning path</strong>, allowing you to build from foundational principles to advanced techniques with logical, structured progression. There’s <strong>executive oversight</strong> of content, which means no redundancy, no gaps, and no guesswork about which course to take next. You can be confident that you receive the same high quality CenterStat training no matter which course or instructor you choose.&nbsp;</p>



<hr class="wp-block-separator has-alpha-channel-opacity"/>



<p><strong>10. A Balanced Focus on Theory and Practice</strong></p>



<p>At CenterStat, we never sacrifice one side of learning for the other. Our courses are carefully designed to strike a <strong>perfect balance between statistical theory and hands-on application</strong>.</p>



<p>You’ll learn the mathematical and conceptual foundations of each model <strong>alongside detailed walkthroughs</strong> of how to apply them to real-world data using modern software tools. This dual focus ensures that your understanding is both deep and immediately actionable. This means you&#8217;ll understand the source and solution to esoteric error messages (e.g., the dreaded <em>psi is non-positive definite</em>) while at the same time being able to tell a meaningful and compelling story about your findings to the reader.&nbsp;</p>



<hr class="wp-block-separator has-alpha-channel-opacity"/>



<p><strong>11. Rigorous Yet Engaging Instructional Style</strong></p>



<p>Too often, statistical education swings between extremes—either overly dry and abstract, or superficial and oversimplified. Our approach is different. Our instructors present complex concepts in a <strong>rigorous yet colloquial style</strong> that makes advanced material feel approachable and—even enjoyable and at times downright humorous (just wait for Dan to start showing pictures of Patrick in awkward stages of childhood and adolescence).</p>



<p>We explain not just what to do, but <em>why</em> it works, using plain language, relevant examples, and a touch of humor to keep you engaged.</p>



<hr class="wp-block-separator has-alpha-channel-opacity"/>



<p><strong>12. Recognized Excellence in Teaching and Research</strong></p>



<p>Many of our instructors are <strong>award-winning educators and researchers</strong>, recognized nationally and internationally for their ability to communicate complex ideas clearly and meaningfully. When you learn from CenterStat, you’re learning from the best—not just in statistics, but in the <strong>art of </strong><strong><em>teaching</em></strong><strong> statistics</strong>. Some platforms allow any instructor to post any material they choose, but at CenterStat we hand pick our instructors from the best in the world and then guide them in developing their materials in the CenterStat Way.&nbsp;</p>



<hr class="wp-block-separator has-alpha-channel-opacity"/>



<p><strong>13. Full Transparency: Detailed Syllabi, Sample Content, and Clear Course Overviews</strong></p>



<p>We understand that enrolling in a course is a commitment. That’s why we provide <strong>clear, detailed information about every workshop</strong> before you register. This includes full syllabi, sample materials, overview videos, clear learning objectives, software covered, and time expectations.&nbsp;</p>



<p>We remove the guesswork so you can make confident, informed decisions about your training.</p>



<hr class="wp-block-separator has-alpha-channel-opacity"/>



<p><strong>14. Tuition Support for Degree-Seeking Students</strong></p>



<p>We’re proud to support the academic journey. If you’re an undergraduate or graduate student currently enrolled in a <strong>degree-granting program</strong>, you may qualify for <strong>reduced tuition rates</strong>.</p>



<p>We believe students shouldn’t have to wait for graduation—or major grant funding—to receive the training they need today. Let CenterStat meet your training goals no matter where you are in your professional development.&nbsp;</p>



<hr class="wp-block-separator has-alpha-channel-opacity"/>



<p><strong>15. Certificates of Completion</strong></p>



<p>Upon completing any CenterStat workshop, you’ll receive a <strong>certificate of completion</strong>—a valuable credential that demonstrates your commitment to professional development. Whether for your CV, academic portfolio, or professional advancement, your certificate signals real achievement in a rigorous and respected program.</p>



<hr class="wp-block-separator has-alpha-channel-opacity"/>



<p><strong>Experience the CenterStat Difference</strong></p>



<p>At a time when online education is more accessible than ever, <strong>quality still matters</strong>. CenterStat stands apart because we combine <strong>deep academic rigor with practical usability</strong>, <strong>affordable pricing with unparalleled value</strong>, and <strong>a learner-first mindset</strong> that informs everything we do.</p>



<p>From our award-winning instructors to our lifetime access, from real-data software demonstrations to our commitment to educational equity, CenterStat doesn’t just deliver courses: we deliver <strong>clarity, confidence, and competence</strong> in statistical methods all at an affordable price.</p>



<p>Ready to elevate your statistical skills with the best in the field?</p>



<p><strong>Explore our workshops today</strong>—and experience for yourself why CenterStat is trusted by students, scholars, and professionals around the world.<strong><br></strong></p>



<p></p>
<p>The post <a href="https://centerstat.org/why-centerstat-is-the-best-choice-for-online-statistics-training/">Why CenterStat is the Best Choice for Online Statistics Training</a> appeared first on <a href="https://centerstat.org">CenterStat</a>.</p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>How CenterStat Workshops Empower Researchers at Every Level</title>
		<link>https://centerstat.org/how-centerstat-workshops-empower-researchers-at-every-level/</link>
		
		<dc:creator><![CDATA[Patrick Curran and Dan Bauer]]></dc:creator>
		<pubDate>Tue, 06 May 2025 17:00:13 +0000</pubDate>
				<category><![CDATA[Announcement]]></category>
		<category><![CDATA[Help Desk]]></category>
		<guid isPermaLink="false">https://centerstat.org/?p=45687</guid>

					<description><![CDATA[<p>In the ever-evolving world of scientific research, having the right analytical tools—and the confidence to use them—is essential. At CenterStat, we understand that researchers face&#8230;</p>
<p>The post <a href="https://centerstat.org/how-centerstat-workshops-empower-researchers-at-every-level/">How CenterStat Workshops Empower Researchers at Every Level</a> appeared first on <a href="https://centerstat.org">CenterStat</a>.</p>
]]></description>
										<content:encoded><![CDATA[
<p>In the ever-evolving world of scientific research, having the right analytical tools—and the confidence to use them—is essential. At CenterStat, we understand that researchers face complex questions that demand rigorous methods and practical, real-world applications. That’s why our workshops are designed not just to educate, but to empower. Whether you&#8217;re a graduate student stepping into your first research project or a seasoned investigator looking to sharpen your quantitative edge, CenterStat training sessions provide the guidance, clarity, and confidence needed to thrive.</p>



<h2 class="wp-block-heading"><strong>Built for All Stages of the Research Journey</strong></h2>



<p>One of the greatest strengths of our workshops is their accessibility across experience levels. We’ve seen participants ranging from doctoral students and postdocs to established professors and NIH-funded investigators. Each of our programs—whether focused on multilevel modeling, structural equation modeling, longitudinal data analysis, or machine learning—is structured to accommodate varying levels of statistical background while still diving deep into theory and application.</p>



<p><em>“The key strength was the clarity of communication and the balance between technicalities and intuition. Dr. Curran &amp; Dr. Bauer did not shy away from technical aspects of the content but also communicated about them as intuitively as possible.”</em><em><br></em> — <strong>Workshop participant, Introduction to Structural Equation Modeling</strong></p>



<h2 class="wp-block-heading"><strong>Practical Skills You Can Use Immediately</strong></h2>



<p>CenterStat workshops are not passive lectures—they’re dynamic, hands-on experiences. Each session blends foundational theory with real data exercises, ensuring that what you learn translates directly to your work. Whether you’re writing a grant, preparing a manuscript, or designing a study, our goal is to give you tools that make an impact now.</p>



<p><em>“Craig is easily one of the best statistics professors I have ever had. He has taken an incredibly complex topic that I have previously struggled to understand and apply to my own data and made it possible for me to feel confident in using and interpreting the analyses on my own</em><em><br></em> — <strong>Workshop participant, Modern Missing Data Analysis</strong></p>



<h2 class="wp-block-heading"><strong>Expert Instructors Who Are Also Practitioners</strong></h2>



<p>Our instructors are internationally recognized scholars who don’t just teach statistics—they use them every day in their own research. They know the challenges of applying quantitative methods in real-life scenarios, and they bring that experience into every lecture and lab session.</p>



<p><em>“This is the most interactive online workshop I have ever attended. It kept me very engaged, motivated to learn, and excited to apply techniques and discuss with the group. Greg and Emily are very knowledgeable, passionate, and eager to share their knowledge with attendees. They are fantastic at answering questions promptly, and they made the entire experience conversational, interactive, and exciting.”</em><em><br></em> — <strong>Workshop participant, Applied Qualitative Research</strong></p>



<h2 class="wp-block-heading"><strong>A Supportive Learning Community</strong></h2>



<p>When you join a CenterStat workshop, you’re not just gaining new knowledge—you’re joining a network of fellow researchers who are equally passionate about improving their craft. Our sessions foster collaboration, discussion, and long-term academic connections.</p>



<p>We also offer ongoing support after the workshop ends, including access to course materials, helpful guides, and recommendations for continuing education.</p>



<h2 class="wp-block-heading"><strong>Results That Speak for Themselves</strong></h2>



<p>CenterStat alumni have gone on to publish in top-tier journals, secure major grants, and teach advanced methods at leading institutions. We take pride in helping researchers unlock their full potential by providing training that’s relevant, rigorous, and rooted in real-world impact.</p>



<p><em>“This is my third C&amp;B workshop and I&#8217;ve really enjoyed all of them. I especially appreciate the attention to the algebra behind models&#8211;they&#8217;ve help me understand other/new-to-me approaches in method papers.”</em><em><br></em> — <strong>Workshop participant, Latent Curve Modeling</strong><strong><br></strong></p>



<h3 class="wp-block-heading"><strong>Ready to Level Up Your Quantitative Skills?</strong></h3>



<p>Explore upcoming workshop opportunities and see how CenterStat can support your research journey at<a href="https://www.centerstat.org/"> www.centerstat.org</a>. Whether you&#8217;re looking to master new techniques or refresh your knowledge, our expert-led sessions will help you gain the confidence and skills to tackle your most important research questions.</p>
<p>The post <a href="https://centerstat.org/how-centerstat-workshops-empower-researchers-at-every-level/">How CenterStat Workshops Empower Researchers at Every Level</a> appeared first on <a href="https://centerstat.org">CenterStat</a>.</p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>CenterStat Announces the Availability of Training Credits</title>
		<link>https://centerstat.org/centerstat-announces-the-availability-of-training-credits/</link>
		
		<dc:creator><![CDATA[Patrick]]></dc:creator>
		<pubDate>Tue, 08 Apr 2025 19:56:09 +0000</pubDate>
				<category><![CDATA[Announcement]]></category>
		<category><![CDATA[Uncategorized]]></category>
		<category><![CDATA[CenterStat Training Credits]]></category>
		<guid isPermaLink="false">https://centerstat.org/?p=45342</guid>

					<description><![CDATA[<p>We are pleased to announce that both individuals and organizations can now purchase CenterStat Training Credits using existing resources that can be applied toward future&#8230;</p>
<p>The post <a href="https://centerstat.org/centerstat-announces-the-availability-of-training-credits/">CenterStat Announces the Availability of Training Credits</a> appeared first on <a href="https://centerstat.org">CenterStat</a>.</p>
]]></description>
										<content:encoded><![CDATA[
<p id="trainingcredits">We are pleased to announce that both individuals and organizations can now purchase <strong>CenterStat Training Credits</strong> using <em>existing </em>resources that can be applied toward <em>future </em>enrollments in any of our workshops.</p>



<p>Credits are denominated in US dollars and are redeemable for up to five years from purchase, so purchasing credits is a great way to convert time-limited funds into an extended opportunity for professional development. Credits are fully transferable to multiple individuals without restrictions, making it easy to provide professional development opportunities to yourself or members of your organization.</p>



<p>As our way of trying to help out in tough times, we will also add an <strong>extra $100</strong> in credits to your account for <strong>every $1000</strong> of training credits you purchase.</p>



<p>Please visit our <a href="https://centerstat.org/trainingcredits/" target="_blank" rel="noreferrer noopener">Training Cr</a><a href="https://centerstat.org/trainingcredits/">edits</a> page for complete details.</p>
<p>The post <a href="https://centerstat.org/centerstat-announces-the-availability-of-training-credits/">CenterStat Announces the Availability of Training Credits</a> appeared first on <a href="https://centerstat.org">CenterStat</a>.</p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>What is a suppressor variable, and how does this differ from confounding and mediation?</title>
		<link>https://centerstat.org/what-is-a-suppressor/</link>
		
		<dc:creator><![CDATA[Patrick]]></dc:creator>
		<pubDate>Fri, 24 Jan 2025 13:53:26 +0000</pubDate>
				<category><![CDATA[Help Desk]]></category>
		<category><![CDATA[Uncategorized]]></category>
		<category><![CDATA[confounding]]></category>
		<category><![CDATA[mediation]]></category>
		<category><![CDATA[suppressor]]></category>
		<category><![CDATA[third variable]]></category>
		<guid isPermaLink="false">https://centerstat.org/?p=43350</guid>

					<description><![CDATA[<p>A constant source of confusion within the multiple regression model (and the general linear model more broadly) relates to the terms suppression and suppressor variable. </p>
<p>The post <a href="https://centerstat.org/what-is-a-suppressor/">What is a suppressor variable, and how does this differ from confounding and mediation?</a> appeared first on <a href="https://centerstat.org">CenterStat</a>.</p>
]]></description>
										<content:encoded><![CDATA[
<p>A constant source of confusion within the multiple regression model (and the general linear model more broadly) relates to the terms <em>suppression</em> and <em>suppressor variable</em>. Indeed, it is not uncommon to see suppression invoked anytime some unanticipated or inexplicable finding is obtained that must be explained away. This is particularly evident when a strongly hypothesized relation is not found: <em>The model results would have supported our hypotheses had they not been obscured by an omitted suppressor variable</em>. What we will see is that (1) this statement is not an entirely accurate use of the term suppression, and (2) suppressor variables can be quite common, easily understood, and wholly accountable by substantive theory. So let&#8217;s think about this a bit more closely, because it really is pretty cool.</p>



<p>To understand suppression, we first need to remind ourselves of a simple two-predictor multiple regression model. Although throughout this note we focus on two predictors and one outcome, all of the concepts easily generalize to multiple predictors, sets of predictors, and even multiple outcomes (e.g., as might be found in a path analysis or structural equation model). This simple two predictor model can be expressed as</p>



<p class="has-text-align-center"><em>y</em>=b<sub>0</sub>+b<sub>1</sub><em>x</em><sub>1</sub>+b<sub>2</sub><em>x</em><sub>2</sub>+r</p>



<p>where<em> y </em>is the outcome, b<sub>0</sub> is the intercept, b<sub>1</sub> and b<sub>2</sub> are the regression coefficients relating<em> x</em><sub>1</sub> and<em> x</em><sub>2</sub> to y, respectively, and r is the residual. As always, each regression coefficient is the unique relation between that predictor and the outcome when controlling for (or <em>above and beyond</em>) the effects of the other predictor. In many situations, if two predictors are correlated with one another then the unique relation of one predictor and the outcome controlling for the other predictor is <em>smaller</em> than that same predictor when considered alone. (Spoiler alert: in suppression the opposite occurs, which is what makes it so incredibly weird).</p>


<div class="wp-block-image">
<figure class="alignleft size-full is-resized"><img decoding="async" width="782" height="469" src="https://centerstat.org/wp-content/uploads/2025/01/fig01-1.png" alt="" class="wp-image-43377" style="width:250px;height:auto" srcset="https://centerstat.org/wp-content/uploads/2025/01/fig01-1.png 782w, https://centerstat.org/wp-content/uploads/2025/01/fig01-1-300x180.png 300w, https://centerstat.org/wp-content/uploads/2025/01/fig01-1-768x461.png 768w, https://centerstat.org/wp-content/uploads/2025/01/fig01-1-624x374.png 624w, https://centerstat.org/wp-content/uploads/2025/01/fig01-1-600x360.png 600w" sizes="(max-width: 782px) 100vw, 782px" /></figure></div>


<p>The typical situation is best seen using a Venn diagram. Consider the bivariate relation between<em> x</em><sub>1</sub> and<em> y </em>where, for the moment, we ignore<em> x</em><sub>2</sub>. In the Venn diagram the bivariate relation between these two variables is denoted as area <em>a</em>. This represents the relation between<em> x</em><sub>1</sub> and <em>y</em> ignoring the effect of<em> x</em><sub>2</sub>. </p>


<div class="wp-block-image">
<figure class="alignright size-full is-resized"><img decoding="async" width="793" height="711" src="https://centerstat.org/wp-content/uploads/2025/01/fig02-1.png" alt="" class="wp-image-43378" style="width:228px;height:auto" srcset="https://centerstat.org/wp-content/uploads/2025/01/fig02-1.png 793w, https://centerstat.org/wp-content/uploads/2025/01/fig02-1-300x269.png 300w, https://centerstat.org/wp-content/uploads/2025/01/fig02-1-768x689.png 768w, https://centerstat.org/wp-content/uploads/2025/01/fig02-1-624x559.png 624w, https://centerstat.org/wp-content/uploads/2025/01/fig02-1-600x538.png 600w" sizes="(max-width: 793px) 100vw, 793px" /></figure></div>


<p>Often, however, we are interested in the unique effects of two (or more) predictors (as well as the joint effect, but we ignore this for now). To obtain these, we bring both predictors into the model at the same time. Typically,<em> x</em><sub>2</sub> is correlated with both<em> x</em><sub>1</sub> and <em>y</em>. As seen in the diagram, in the presence of the second predictor, the effect of the first predictor, <em>a</em>, is usually smaller than it was before. This is quite natural in that the part of<em> x</em><sub>1</sub> that is shared with<em> x</em><sub>2</sub> is removed when assessing the unique relation between<em> x</em><sub>1</sub> and<em> y</em>, making <em>a</em> smaller. This is business as usual.</p>



<p>However, let&#8217;s make things a bit stranger. More than 80 years ago, a brilliant quantitative psychologist named Paul Horst found a situation in which the relation between<em> x</em><sub>1</sub> and<em> y </em>was actually <em>larger </em>in the presence of<em> x</em><sub>2</sub> than when assessed in the absence of<em> x</em><sub>2</sub> (Horst, 1941). Considering the above Venn diagrams, this makes absolutely no sense at all; it actually seems downright impossible. Yet it most definitely exists, and Horst somewhat unfortunately termed this situation <em>suppression</em>. It is unfortunate because<em> x</em><sub>2</sub> is not suppressing the relation between<em> x</em><sub>1</sub> and<em> y </em>(which many researchers assume). In actuality,<em> x</em><sub>2</sub> is suppressing irrelevant variance in<em> x</em><sub>1</sub> and, by doing so, <em>enhances</em> the relation between<em> x</em><sub>1</sub> and y. It might have been better to refer to<em> x</em><sub>2</sub> as an &#8220;enhancer&#8221; rather than a &#8220;suppressor&#8221;, but that historical ship has sailed so we are stuck with this terminology.</p>



<p>To understand suppression better, let&#8217;s first think about the substantive application in which Horst observed this phenomenon. He was part of a research team evaluating pilots during World War II. The pilots completed paper-and-pencil assessments measuring three types of cognitive reasoning: mechanical, numerical, and spatial. These three measures were then used to predict a score representing piloting ability. However, these measures were not as strongly predictive of piloting ability as had been anticipated. A fourth predictor was then included that was a measure of general verbal ability. Verbal ability was correlated with each of the three reasoning measures (as would be expected), but was <em>not</em> correlated with piloting ability (as would also be expected). Consistent with expectations, when verbal ability was added to the regression model it did not uniquely predict piloting ability. However, quite unexpectedly, the effects of all three reasoning measures were markedly <em>larger</em> compared to the model in which verbal ability was omitted. Horst determined that the reason was that verbal ability was removing irrelevant information from the three reasoning measures (that is, <em>suppressing</em> the part of the variance in reasoning that was related to verbal ability but unrelated to piloting ability) and this in turn <em>enhanced</em> the relations between what was left over in the reasoning measures and pilot ability, i.e., the unique relations.</p>



<p>It is helpful to consider a simple hypothetical example. We will define<em> x</em><sub>1</sub> to be the predictor variable of interest (say mechanical reasoning) and<em> x</em><sub>2</sub> the suppressor variable (say verbal ability). The simplest pattern of correlations consistent with traditional suppression is when<em> x</em><sub>1</sub> is correlated with <em>y</em>,<em> x</em><sub>1</sub> is correlated with<em> x</em><sub>2</sub>, and<em> x</em><sub>2</sub> is not correlated with y. Of course, in any sample data there will rarely be a zero correlation between the suppressor and the outcome, but it might still obtain some negligible value. Let&#8217;s further say that the correlation between<em> x</em><sub>1</sub> and<em> y </em>is .25, between<em> x</em><sub>1</sub> and<em> x</em><sub>2</sub> is .70, and between<em> x</em><sub>2</sub> and<em> y </em>is zero. If we consider a model in which only<em> x</em><sub>1</sub> predicts <em>y</em>, the standardized regression coefficient for<em> x</em><sub>1</sub> is equal to .25 with a squared semi-partial correlation of .06 (that is,<em> x</em><sub>1</sub> uniquely accounts for 6% of the variance in <em>y</em>). However, if we add<em> x</em><sub>2</sub> as a second predictor, the standardized regression coefficient for<em> x</em><sub>1</sub> increases to .49 and the squared semi-partial correlation doubles to .12 (that is,<em> x</em><sub>1</sub> now uniquely accounts for 12% of the variance in <em>y</em>). The unique effect of <em>x</em><sub>1</sub> on <em>y</em> when controlling for <em>x</em><sub>2</sub> is markedly stronger than the bivariate relation of <em>x</em><sub>1</sub> with <em>y</em>, the hallmark of suppression.</p>



<p>What on Earth is going on? Of course, the presence or absence of the suppressor does not change the bivariate relation between the predictor and the outcome (the correlation between<em> x</em><sub>1</sub> and<em> y </em>is always .25). However, what the suppressor does change is the unique variability in the predictor that is available to be related to the outcome.</p>


<div class="wp-block-image">
<figure class="alignright size-full is-resized"><img loading="lazy" decoding="async" width="887" height="729" src="https://centerstat.org/wp-content/uploads/2025/01/fig03-1.png" alt="" class="wp-image-43379" style="width:297px;height:auto" srcset="https://centerstat.org/wp-content/uploads/2025/01/fig03-1.png 887w, https://centerstat.org/wp-content/uploads/2025/01/fig03-1-300x247.png 300w, https://centerstat.org/wp-content/uploads/2025/01/fig03-1-768x631.png 768w, https://centerstat.org/wp-content/uploads/2025/01/fig03-1-624x513.png 624w, https://centerstat.org/wp-content/uploads/2025/01/fig03-1-600x493.png 600w" sizes="auto, (max-width: 887px) 100vw, 887px" /></figure></div>


<p>We can see this in a simple re-arrangement of the Venn diagram (the Venn diagram is an imperfect representation of the underlying mathematics, but visually gives a sense what is happening here). This shows that the suppressor, <em>x</em><sub>2</sub>, correlates with the predictor, <em>x</em><sub>1</sub>, as indicated by area <em>b</em>, but does not correlate with the outcome (reflected in the lack of overlap of the circles for <em>x</em><sub>2</sub> and <em>y </em>). Further, we can see that controlling for the suppressor reduces (or <em>suppresses</em>) part of the variance in<em> x</em><sub>1</sub> that is unrelated to<em> y </em>(area <em>b</em>). This, in turns, <em>enhances</em> the proportional relation between<em> x</em><sub>1</sub> and<em> y </em>(represented by area <em>a</em>). This is the core of suppression.</p>



<p>There have been dozens of papers written on suppression following Horst&#8217;s initial discovery, and we note several of these below. Many of these propose specific subtypes of suppression and describe under what unique conditions these might be encountered in practice. However, a concise general definition was given by Conger (1974) who wrote &#8220;<em>A suppressor variable is defined to be a variable which increases the predictive validity of another variable (or set of variables) by its inclusion in a regression equation. This variable is a suppressor only for those variables whose regression weights are increased</em>.&#8221; Importantly, this means that a variable is not inherently a suppressor in and of itself. Instead, a suppressor is defined by the impact it has on <em>other</em> variables in the model. That is, a variable might be a suppressor in one model but not in another.</p>



<p>This brings us to two initial points. First, there is nothing about suppression that is magical, mysterious, or misunderstood. The papers noted below explain in gory detail exactly what suppression is and under what conditions it exists, so be suspect of a paper that says &#8220;Suppression is a long-misunderstood issue&#8230;&#8221;. It is not. Initially confusing? Yes. Misunderstood? No.</p>



<p>Second, suppression is not some unavoidable artifact of measurement or estimation but instead can be fully accounted by substantive theory. Horst&#8217;s example is just one of many in the literature, nearly all of which make perfect sense within a given theoretical framework. As such, it is often beneficial to think about potential suppressors during the design phase of a study so that all relevant variables can be included in the analysis. </p>



<p>However, our third and final point is a bit of a punch in the face: we must consider two additional competing explanations for the role of a third variable in our models: confounding and mediation.</p>



<p>By far the clearest treatment of this was given by MacKinnon, Krull and Lockwood (2000) in a title that could not be more on point: <em>Equivalence of the Mediation, Confounding, and Suppression Effect</em>. You nearly don&#8217;t have to read the paper given the title. The paper opens with, &#8220;<em>Once a relationship between two variables has been established, it is common for researchers to consider the role of a third variable in this relationship</em>.&#8221; This is precisely what we have considered thus far. But to better see this point, we move from Venn diagrams to path diagrams. Let&#8217;s first consider the two-predictor regression that we have discussed up to this point:</p>


<div class="wp-block-image">
<figure class="aligncenter size-full is-resized"><img loading="lazy" decoding="async" width="986" height="476" src="https://centerstat.org/wp-content/uploads/2025/01/fig04-1.png" alt="" class="wp-image-43380" style="width:330px;height:auto" srcset="https://centerstat.org/wp-content/uploads/2025/01/fig04-1.png 986w, https://centerstat.org/wp-content/uploads/2025/01/fig04-1-300x145.png 300w, https://centerstat.org/wp-content/uploads/2025/01/fig04-1-768x371.png 768w, https://centerstat.org/wp-content/uploads/2025/01/fig04-1-624x301.png 624w, https://centerstat.org/wp-content/uploads/2025/01/fig04-1-600x290.png 600w" sizes="auto, (max-width: 986px) 100vw, 986px" /></figure></div>


<p>This shows the usual expression of two correlated predictors and one outcome. Note that there are three measured variables, and these are all related to one another (the curved arrow reflects the correlation between the two predictors, and the two one-headed arrows reflect the partial regression coefficients). In a suppression situation, <em> x</em><sub>1</sub> <em>enhances</em> the relation between<em> x</em><sub>1</sub> and y.</p>



<p>However, with a simple re-arrangement of the diagram we get what is called <em>confounding</em>:</p>


<div class="wp-block-image">
<figure class="aligncenter size-full is-resized"><img loading="lazy" decoding="async" width="977" height="716" src="https://centerstat.org/wp-content/uploads/2025/01/fig05-1.png" alt="" class="wp-image-43381" style="width:254px;height:auto" srcset="https://centerstat.org/wp-content/uploads/2025/01/fig05-1.png 977w, https://centerstat.org/wp-content/uploads/2025/01/fig05-1-300x220.png 300w, https://centerstat.org/wp-content/uploads/2025/01/fig05-1-768x563.png 768w, https://centerstat.org/wp-content/uploads/2025/01/fig05-1-624x457.png 624w, https://centerstat.org/wp-content/uploads/2025/01/fig05-1-600x440.png 600w" sizes="auto, (max-width: 977px) 100vw, 977px" /></figure></div>


<p>Note that all we have done is changed the correlation between the two predictors to a regression coefficient and now<em> x</em><sub>2</sub> is a <em>confounder</em> in that it predicts both<em> x</em><sub>1</sub> and <em>y</em>. This is the situation that is so fun to teach because we can give examples such as the number of fire trucks sent to a fire is positively correlated to the amount of damage done at the fire; but when the confounder of severity of fire is included, there is no relation between number of trucks and damage. Importantly, whereas a suppressor <em>enhances</em> the relation between<em> x</em><sub>1</sub> and <em>y</em>, including a confounder as a second regressor <em>reduces</em> this same relation.</p>



<p>Finally, re-directing one arrow in the above path diagram results in <em>mediation</em>:</p>


<div class="wp-block-image">
<figure class="aligncenter size-full is-resized"><img loading="lazy" decoding="async" width="1163" height="540" src="https://centerstat.org/wp-content/uploads/2025/01/fig06-1.png" alt="" class="wp-image-43382" style="width:309px;height:auto" srcset="https://centerstat.org/wp-content/uploads/2025/01/fig06-1.png 1163w, https://centerstat.org/wp-content/uploads/2025/01/fig06-1-300x139.png 300w, https://centerstat.org/wp-content/uploads/2025/01/fig06-1-1024x475.png 1024w, https://centerstat.org/wp-content/uploads/2025/01/fig06-1-768x357.png 768w, https://centerstat.org/wp-content/uploads/2025/01/fig06-1-624x290.png 624w, https://centerstat.org/wp-content/uploads/2025/01/fig06-1-600x279.png 600w" sizes="auto, (max-width: 1163px) 100vw, 1163px" /></figure></div>


<p>Now<em> x</em><sub>2</sub> <em>explains</em> the relation between<em> x</em><sub>1</sub> and <em>y</em>. For example, the predictor might be parent&#8217;s alcohol use, the outcome is the child&#8217;s alcohol use, and the mediator is impaired parenting. The inference is that the parent&#8217;s alcohol use impairs their own parenting behavior, and this in turn increases the probability that the child will drink alcohol themselves. In sum, suppression <em>enhances</em> the relation between a predictor and the outcome, confounding <em>reduces</em> the relation, and mediation <em>explains</em> the relation. How the heck do we differentiate among the three? MacKinnon et al. (2000) argue that you do not, and they demonstrate that the statistical tests of these three effects are all identical: each model is a simple re-expression of the others. They conclude the paper saying, &#8220;<em>The statistical procedures provide no indication of which type of effect is being tested. That information must come from other sources</em>.&#8221; The &#8220;other sources&#8221; to which they refer are prior knowledge and theory. All three effects are statistically isomorphic, and only theory can discern which most likely holds in the population. Further differentiation might also be possible by moving to experimental or longitudinal designs that allow for better testing of hypotheses about causal pathways.</p>



<div class="wp-block-uagb-separator uagb-block-8c58b629"><div class="wp-block-uagb-separator__inner" style="--my-background-image:"></div></div>



<h4 class="wp-block-heading">Suggested Readings</h4>



<p>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp; Conger A.J. (1974). A revised definition for suppressor variables: A guide to their identification and interpretation. <em>Educational Psychological Measurement, 34</em>, 35–46.</p>



<p>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp; Horst P. (1941). The role of predictor variables which are independent of the criterion. Social Science Research Council Bulletin, 48, 431–436</p>



<p>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp; MacKinnon, D. P., Krull, J. L., &amp; Lockwood, C. M. (2000). Equivalence of the mediation, confounding and suppression effect. <em>Prevention Science, 1</em>, 173-181.</p>



<p>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp; Tzelgov, J., &amp; Henik, A. (1991). Suppression situations in psychological research: Definitions, implications, and applications. <em>Psychological Bulletin, 109</em>, 524-536.</p>



<p>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp; Velicer W.F. (1978). Suppressor variables and the semipartial correlation coefficient. <em>Educational and Psychological Measurement, 38</em>, 953–958</p>
<p>The post <a href="https://centerstat.org/what-is-a-suppressor/">What is a suppressor variable, and how does this differ from confounding and mediation?</a> appeared first on <a href="https://centerstat.org">CenterStat</a>.</p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>What are ROC curves and how are these used to aid decision making?</title>
		<link>https://centerstat.org/what-are-roc-curves-and-how-are-these-used-to-aid-decision-making/</link>
		
		<dc:creator><![CDATA[Patrick Curran and Dan Bauer]]></dc:creator>
		<pubDate>Mon, 18 Sep 2023 21:18:41 +0000</pubDate>
				<category><![CDATA[Help Desk]]></category>
		<category><![CDATA[Uncategorized]]></category>
		<guid isPermaLink="false">https://centerstat.local/?p=35615</guid>

					<description><![CDATA[<p>ROC curves are used to identify cut points on continuous measures at which to make binary decisions. An example is a depression screener used to identify people who may qualify for a major depression diagnosis.  This post explains the concepts behind ROC curves and how they are used for decision making.</p>
<p>The post <a href="https://centerstat.org/what-are-roc-curves-and-how-are-these-used-to-aid-decision-making/">What are ROC curves and how are these used to aid decision making?</a> appeared first on <a href="https://centerstat.org">CenterStat</a>.</p>
]]></description>
										<content:encoded><![CDATA[
<p>One of the most vexing challenges in all of statistics is the need to make a valid and reliable probabilistic assessment about some unknown condition or state of affairs based solely on information gathered from sample data. Indeed, this is the foundation of traditional null hypothesis testing: there is some unknown condition in the population (the null hypothesis is either true or false) and we must decide whether we <em>believe</em> the null hypothesis is true or false based solely on our sample data. There are thus four possible outcomes in which two decisions are correct (we reject the null when the null is false or we retain the null when the null is true) and two are incorrect (we reject the null when it is actually true or we retain the null when it is actually false). Type I Error is the probability of rejecting a null that is true, and Type II error is the probability of retaining a null that is false. This is commonly displayed in a traditional two-by-two table is the cornerstone of a frequentist hypothesis testing strategy that we all encounter every single day in our work.</p>


<div class="wp-block-image">
<figure class="aligncenter size-full"><img loading="lazy" decoding="async" width="401" height="136" src="https://centerstat.org/wp-content/uploads/2023/09/image-3.png" alt="" class="wp-image-35625" srcset="https://centerstat.org/wp-content/uploads/2023/09/image-3.png 401w, https://centerstat.org/wp-content/uploads/2023/09/image-3-300x102.png 300w" sizes="auto, (max-width: 401px) 100vw, 401px" /></figure></div>


<p> </p>



<p>This same approach to decision making appears routinely in daily life. For example, we might want to know if a person referred to a clinic is likely to be diagnosed with major depression or not: they either are truly depressed or not (which is unknown to us) and we must decide based on some brief screening test if we believe it is likely that they suffer from depression. Or we might want to know if it is likely someone has a medical diagnosis that requires a more invasive biopsy procedure. Or, one that is near and dear to us all, we may want to know if we do or do not have COVID. There is a “true” condition (you really do or really do not have COVID) and we obtain a positive or negative result on a rapid test we bought for ten dollars from Walgreens. We have precisely the same four possible outcomes as shown above, two that are correct (it says you have COVID if you really do, or you do not have COVID if you really do not) and two that are incorrect (it says you have COVID when you really don&#8217;t, or you do not have COVID if you really do).</p>


<div class="wp-block-image">
<figure class="aligncenter size-full"><img loading="lazy" decoding="async" width="416" height="117" src="https://centerstat.org/wp-content/uploads/2023/09/image-4.png" alt="" class="wp-image-35626" srcset="https://centerstat.org/wp-content/uploads/2023/09/image-4.png 416w, https://centerstat.org/wp-content/uploads/2023/09/image-4-300x84.png 300w" sizes="auto, (max-width: 416px) 100vw, 416px" /></figure></div>


<p> </p>



<p>This is such an important concept that there are specific terms that capture these possible outcomes. <em>Sensitivity</em> is the probability that you will receive a positive rapid test result if you truly have COVID (the probability of a <em>true positive</em> for those with the disease). <em>Specificity</em> is the probability that you will receive a negative rapid test result if you truly do not have COVID (the probability of a <em>true negative</em> for those without the disease). One minus sensitivity thus represents the error of obtaining a false negative result, and one minus specificity represents the error of obtaining a false positive result. However, all of the above assumes that the rapid test has one of two outcomes: the little window on the rapid test either indicates a negative or a positive result. But this begs the very important question, <em>how does the test <u>know</u>?</em> That is, the test is based on a <em>continuous</em> reading of antigen levels yet some scientist in some lab decided that an exact point on this continuous antigen scale would change the test result from negative to positive. Determining this ideal point on a continuum is one of the many uses for ROC analysis.</p>



<p>ROC stands for <em>receiver operating characteristic</em>, the history of which can be traced back to the development of radar during World War II. Radar was in its infancy and engineers were struggling to determine how it could best be calibrated to maximize the probability of identifying a real threat (an enemy bomber, or a <em>true positive</em>) while minimizing the probability of a false alarm (a bird or a rain squall, or a <em>false positive</em>). The challenge was where to best set the continuous sensitivity of the receiver (called <em>gain</em>) to optimally balance these two outcomes. In other words, there was an infinite continuum of possible gain settings and they needed to determine a specific value that would balance true versus false readings. This is precisely the situation in which we find ourselves when using a brief screening instrument to identify depression or blood antigen levels to identify COVID.</p>



<p>To be more concrete, say that we had a 20-item screening instrument for major depression designed to assess whether an individual should be referred for treatment or not, but we don’t know at what specific score a referral should be made. We thus want to examine the ability of the continuous measure to optimally discriminate between true and false positive decisions across all possible cut-offs on the continuum. To accomplish this, we gather a sample of individuals with whom we conduct a comprehensive diagnostic workup to determine “true” depression, and we give the same individuals our brief 20-item screener and obtain a person-specific scale score that is continuously distributed. We can now construct what is commonly called a <em>ROC curve</em> that plots the true positive rate (or sensitivity) against the false positive rate (or one minus specificity) across all possible cut-points on a continuous measure. That is, we can determine how every possible cut-point on the screener discriminates between those who did or did not receive a comprehensive depression diagnosis.</p>



<p>To construct a ROC curve, we begin by creating a bivariate plot in which the <em>y-</em>axis represents <em>sensitivity</em> (or <em>true positives</em>) and the <em>x-</em>axis represents one minus <em>specificity</em> (or <em>false positives</em>). Because we are working in the metric of probabilities each axis is scaled between zero and one. We are thus plotting the true positive rate against the false positive rate across the continuum of possible cut points on the screener. Next, a 45-degree line is fixed from the origin (or 0,0 point) to the upper right quadrant (or 1,1 point) to indicate random discrimination; that is, for a given cut-point on the continuous measure, you are as likely to make a true positive as you are a false positive. However, the key information in the ROC curve is superimposing the sample-based curve that is associated with your continuous screener; this reflects the actual true-vs-false positive rate across all possible cut-offs of your screener. An idealized ROC curve (drawn from Wikipedia) is presented below.</p>


<div class="wp-block-image">
<figure class="aligncenter size-medium"><img loading="lazy" decoding="async" width="300" height="295" src="https://centerstat.org/wp-content/uploads/2023/09/image-2-300x295.png" alt="" class="wp-image-35618" srcset="https://centerstat.org/wp-content/uploads/2023/09/image-2-300x295.png 300w, https://centerstat.org/wp-content/uploads/2023/09/image-2-32x32.png 32w, https://centerstat.org/wp-content/uploads/2023/09/image-2.png 453w" sizes="auto, (max-width: 300px) 100vw, 300px" /></figure></div>


<p> </p>



<p>If the screener has no ability to discriminate between the two groups, the sample-based ROC curve will fall on the 45-degree line. However, that rarely happens in practice; instead, the curve capturing the true-to-false positive rates across all possible cut-points will lie above the 45-degree line indicating that the test is performing better than chance alone. The further the ROC curve deviates from the 45-degree line, the better able the screener is to correctly assign individuals to groups. At the extreme, a perfect screener will fall in the upper left corner (the 0,1 point) indicating all decisions are true positives and none are false positives. This too rarely if ever occurs in practice, and a screener will nearly always fall somewhere in the upper-left area of the plot.</p>



<p>But how do we know if our sample-based curve is meaningfully higher than the 45-degree line? There are many ways that have been proposed to evaluate this, but the most common is computing the <em>area under the curve</em>, or AUC. Because the plot defines a unit square (that is, it is one unit wide and one unit tall), 50% of the area of the square falls below the 45-degree line. Because we are working with probabilities, we can literally interpret this to mean that a there is a 50-50 chance a randomly drawn person from the depressed group has a higher score on the screener than a randomly drawn person from the non-depressed group. This of course reflects that the screener has no better than random chance of correctly specifying an individual. But what if the AUC for the screener was say .80? This would reflect that there is a probability of .8 that a randomly drawn person from the depressed group will have a higher score on the screener than a randomly drawn person from the non-depressed group. In other words, the screener is able to <em>discriminate</em> between the two groups at a higher rate than chance alone. But how high is high enough? There is not really a “right” answer, but conventional benchmarks are that AUCs over .90 are “excellent”, values between .70 and .90 are “acceptable” and values below .70 are “poor”. &nbsp;Like most general benchmarks in statistics, these are subjective, and much will ultimately depend on the specific theoretical question, measures, and sample at hand. (We could also plot multiple curves to compare two or more screeners, but we don’t detail this here.)</p>



<p>Note, however, that the AUC is a characteristic of the screener itself and we have not yet determined the optimal cut-point to use to classify individual cases. For example, say we wanted to determine the optimal value on our 20-item depression screener that would maximize the true positives and minimize the false positives in our referral for individuals to obtain a comprehensive diagnostic evaluation. Imagine that individual scores could range in value from zero to 50 and we could in principle set the cut-off value at any point on the scale. The ROC curve allows us to compare the true positive to false positive rate across the entire range of the screener and estimate what the true vs. false positive classification at each and every value of the screener. We then can select the optimal value that best balances true positives from false positives, and that value becomes are cut-off point at which we demarcate who is referred for a comprehensive diagnostic evaluation and those who are not. There are a variety of methods for accomplishing this goal, including computing the closest point at which the curve approaches the upper-left corner, the point at which a certain ratio of true-to-false positives is reached, and using more recent methods drawn from Bayesian estimation and machine learning. Some of these methods become quite complex, and we do not detail these here.</p>



<p>Regardless of method used, it is important to realize that the optimal cut-point may not be universal but varies by one or more moderators (e.g., biological sex or age) such that one cut-point is ideal for children and another for adolescents. Further, the ideal cut-point might be informed by the relative cost of making a true vs. false positive. For example, a more innocuous example might be determining if a child might benefit from additional tutoring in mathematics compared to a much more severe determination of whether an individual might suffer from severe depression and be at risk for self-harm. Different criteria might be used in determining the optimal cut-point for the former vs. the latter. Importantly, this statistical architecture is quite general and can be applied across a wide array of settings within the social sciences and offers a rigorous and principled method to help guide optimal decision making. We offer several suggested readings below.</p>



<div class="wp-block-uagb-separator uagb-block-8c58b629"><div class="wp-block-uagb-separator__inner" style="--my-background-image:"></div></div>



<h4 class="wp-block-heading">Suggested Readings</h4>



<p>Fan, J., Upadhye, S., &amp; Worster, A. (2006). Understanding receiver operating characteristic (ROC) curves. <em>Canadian Journal of Emergency Medicine, 8</em>, 19-20.</p>



<p>Hart, P. D. (2016). Receiver operating characteristic (ROC) curve analysis: A tutorial using body mass index (BMI) as a measure of obesity. <em>J Phys Act Res</em>, 1, 5-8.</p>



<p>Janssens, A. C. J., &amp; Martens, F. K. (2020). Reflection on modern methods: Revisiting the area under the ROC Curve. <em>International journal of epidemiology, 49</em>, 1397-1403.</p>



<p>Mandrekar, J. N. (2010). Receiver operating characteristic curve in diagnostic test assessment. <em>Journal of Thoracic Oncology, 5</em>, 1315-1316.</p>



<p>Petscher, Y. M., Schatschneider, C., &amp; Compton, D. L. (Eds.). (2013). <em>Applied quantitative analysis in education and the social sciences</em>. Routledge.</p>



<p>Youngstrom, E. A. (2014). A primer on receiver operating characteristic analysis and diagnostic efficiency statistics for pediatric psychology: we are ready to ROC. <em>Journal of Pediatric Psychology, 39</em>, 204-221.</p>
<p>The post <a href="https://centerstat.org/what-are-roc-curves-and-how-are-these-used-to-aid-decision-making/">What are ROC curves and how are these used to aid decision making?</a> appeared first on <a href="https://centerstat.org">CenterStat</a>.</p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>What exactly qualifies as intensive longitudinal data and why am I not able to use more traditional growth models to study stability and change over time?</title>
		<link>https://centerstat.org/ild-help-desk/</link>
		
		<dc:creator><![CDATA[Patrick Curran and Dan Bauer]]></dc:creator>
		<pubDate>Fri, 14 Oct 2022 19:15:04 +0000</pubDate>
				<category><![CDATA[Help Desk]]></category>
		<guid isPermaLink="false">https://centerstat.local/?p=27955</guid>

					<description><![CDATA[<p>This post considers the unique features of intensive longitudinal data (ILD) relative to other more traditional data structures and how we can appropriately analyze ILD given these features</p>
<p>The post <a href="https://centerstat.org/ild-help-desk/">What exactly qualifies as intensive longitudinal data and why am I not able to use more traditional growth models to study stability and change over time?</a> appeared first on <a href="https://centerstat.org">CenterStat</a>.</p>
]]></description>
										<content:encoded><![CDATA[
<p>Recent years have seen increasing interest in the collection and analysis of intensive longitudinal data (or ILD) to generate unique insights into within-person processes and change over time. In this post, we first define ILD by contrasting it to data obtained from other common longitudinal designs. Next, we consider the distinct features of ILD that we must address and can leverage in our analyses. Last, we describe two common approaches for analyzing ILD: multilevel modeling and dynamic structural equation modeling.</p>



<p>To start, there is often much confusion over what constitutes intensive longitudinal data (or ILD), in large part because there exists no formal definition that separates ILD from other types of longitudinal data. That said, ILD tends to fall between two traditional data structures obtained from alternative designs: panel data and time series data. It’s useful to first consider these traditional structures to see how several of their features will combine within ILD.</p>



<p>Historically, the most common method for gathering longitudinal data in psychology and the social and health sciences has been the <em>panel design</em>. Typically, a panel design involves assessing a large sample of subjects (say 200 or more) at a much smaller number of time points (say three to six) that tend to be widely spaced in time (say six or 12 months or more). Panel data are often used to empirically examine long-term trajectories of change that might span multiple years, and common analytic methods include the standard latent curve model or a multilevel growth model. (See our <a href="https://centerstat.org/mlm-v-sem-growth/" target="_blank" rel="noreferrer noopener">prior Help Desk</a> entry on the relation between the LCM and MLM).</p>



<p>A second type of longitudinal design, commonly used in economics among other areas, is the <em>time series</em> <em>design</em>, which resides at the opposite end of the continuum from the panel design. More specifically, a time series design is often based on just a single unit that is repeatedly assessed a very large number of times (say 100 to 200 or more) at intervals that tend to be close together in time (say daily or even hourly). Time series data are often used to empirically examine short-term dynamic processes that might unfold hour-by-hour or day-by-day (e.g., the daily closing cost of the S&amp;P500) and many specialized analytic methods exist to fit models to these highly dense data.</p>



<p>ILD tends to fall between the two extremes of panel data on one end and time series on the other. More specifically, ILD tends to have fewer subjects than panel data but more than time series (say 50 or 100 subjects) and more time points than panel data but possibly fewer than time series (say 30 or 40 assessments). Data might be captured using wearable technology (e.g., heart rate or blood pressure monitors) or by sending random prompts throughout the day via smart phones or other electronic devices (e.g., a tone sounds on a smart phone three times throughout the day and an individual is prompted to respond to a brief feelings survey). As a hypothetical example, a study might be designed to randomly measure nicotine cravings and cigarette use in a sample of 50 individuals four times per day for a two week period resulting in 56 assessments on each individual, thus falling between traditional panel and time series designs in structure.</p>



<p>In the spirit of <em>be careful what you ask for</em>, once you obtain intensive longitudinal data you must then select an optimal modeling strategy to test your motivating hypotheses, and this is not always an easy task. To begin, some longitudinal models that we are familiar with from panel data simply will not work with ILD. Consider the latent curve model (LCM): because the LCM is embedded within the structural equation model, each observed time point is represented by a manifest variable in the model. This works well if the model is fit to annual assessments of some outcome (say antisocial behavior at age 6, 7, 8, 9 and 10) where each age-specific measure serves as an indicator on the underlying latent curve factor. However, the LCM rapidly breaks down with higher numbers of repeated measures in which only one observation may have been obtained at any given assessment (e.g., 9:15am, 9:52am, and so on). For our prior example with 56 repeated assessments taken on 50 subjects, the LCM is simply not an option.</p>



<p>We can next consider the multilevel model (MLM) and it turns out that this option works quite well for many ILD research applications. (See our Office Hours channel on YouTube for a <a href="https://youtu.be/k1TuG5Zxvro" target="_blank" rel="noreferrer noopener">lecture on the MLM </a>with repeated measures data). The MLM approaches the complex ILD structure as nested data in which repeated assessments are nested within individual. Interestingly, unlike the standard LCM, the MLM can be applied to both more traditional panel data and to ILD. The reason is that, whereas the LCM incorporates the passage of time into the factor loading matrix and requires an observed variable at each assessment, the MLM incorporates the passage of time as a numerical predictor in the regression model. As such, the MLM can easily allow for highly dense (meaning many time points) and highly sparse (meaning few or even one assessment is shared by any individual at any given time point) data without problem. (The LCM can under certain circumstances be contorted to accommodate some of these features as well, but the MLM does this seamlessly). However, there are several complications that must be addressed when fitting an MLM to intensive longitudinal data that do not commonly arise in panel data.</p>



<p>The first issue is what is called <em>serial correlation</em> of the residuals for the repeatedly measured outcome.&nbsp; With apologies for the technical terminology, this means is that for a given person, when there is a “bump” at one timepoint, that tends to carry over to the next time point too.&nbsp; For instance, say a person’s average heart rate is 72 BPM. I measured this person at 9:10am and 9:26am. What I don’t know is that this person was late for their 9:00am job, which lead them to move faster and increased their stress, and they had only just arrived at 9:10am.&nbsp; This manifested in a heart rate of 91 BPM at 9:10 and 83 BPM at 9:26. The initial bump has thus not entirely dissipated by the second assessment.</p>



<p>Serial correlation is often not of importance in panel data because these perturbations have long since washed out (the residual correlation goes to zero over the long lags). A person’s heart rate might be higher than usual when I assess them at age 26 because they had a second shot of espresso or got in an argument with a colleague at work, but the effect of the espresso or argument has long since worn off by the time I reassess them at age 27.&nbsp; Of course, even with panel data the repeated measures are correlated, but not because of serial correlation of within-person <em>residuals</em> but because of individual differences in level and change over time.&nbsp; For instance, some people have consistently higher heart rates and others have consistently lower heart rates and this stability will lead to across-person positive correlations in repeated measures. We typically model these individual differences in level and change via latent growth factors / random effects when fitting LCMs / MLMs. Such individual differences may be an important source of correlation in ILD too, but we also have to contend with the serially correlated residuals. Although an added complexity, the MLM is quite well suited at incorporating serial correlations such as these. Complex error terms can be defined among the time-specific residuals such as auto-regressive, Gaussian decay, spatial power, or Toeplitz structures. It is very important these serial correlations be represented in the model if needed both to gain insights into the phenomenon under study and to ensure that other parameter estimates of interest are not biased.</p>



<p>A second issue that often arises in ILD is the presence of cycles or transition points that might occur during the assessment period. For example, daily measures taken over several weeks may vary as a function of weekday vs. weekend (e.g., if studying college drinking) or might cycle regularly throughout a day (e.g., hourly heart rate data varying as a function of waking to sleeping and back to waking). Although such cycles and transition points might be present in panel data as well, these are less likely to occur because there are typically fewer time-linked assessments and these tend to aggregate over longer durations (e.g., if we ask “over the past 30 days” to obtain monthly alcohol use levels, these ratings will implicitly smooth over weekday-weekend differences in daily alcohol use). In contrast, multiple cycles might be observed in ILD spanning a 50 or 60 time point series.</p>



<p>Finally, a third issue is the distinction between within- versus between-person effects. Often ILD is collected with the idea of assessing processes as they unfold in real time for individual participants (“life as lived”). For instance, we might be interested in using ILD to test a negative reinforcement hypothesis for alcohol use. That is, we wish to test the proposition that people drink more than they typically do when they are experiencing increased negative affect under the expectation that this will reduce their negative affect. Using a daily diary study, we measure negative affect each day and alcohol use each night and we build a model to predict alcohol use from negative affect. To fully assess the negative reinforcement hypothesis, we must differentiate the within-person effect (e.g., when my negative affect is higher than usual I drink more than is typical for me)&nbsp; from any between-person correlation that may also exist (e.g., &nbsp;that people who have higher negative affect in general tend to drink more in general).&nbsp; Fortunately, with the MLM we have well developed methods for separating within- and between-person effects, although there are some complications to consider (see our prior <a href="https://centerstat.org/biased-between-group-effects/" target="_blank" rel="noreferrer noopener">help desk</a> post specifically on this issue)</p>



<p>The MLM is thus well suited to address all of these complexities that commonly arise in intensive longitudinal data. Once incorporated, the MLM offers many of the very same advantages as when applied to panel data: time-varying predictors can be incorporated at level-1 with either fixed or random effects, time invariant predictors can be incorporated at level-2, and interactions can be estimated within or across levels of analysis. However, there are two key limitations of the MLM that may or may not arise in a given application. The first is that, similar to the traditional general linear model, the MLM assumes all measures are error-free and all observed variance is &#8220;true&#8221; variance. This is often (if not always) an unrealistic assumption and violation of this assumption can lead to significant biases in the estimated results. The second is that the MLM only allows for one dependent variable at a time and is thus limited to the estimation of unidirectional effects. Say that you are interested in testing the reciprocal relations between depression during the day and substance use that evening, and you obtain multiple daily measures spanning a week of time. The MLM allows for the estimation of the prediction of substance use from depression, but not the simultaneous estimation of the reciprocal prediction of depression from substance use. As such, the MLM is only evaluating one part of the research hypotheses at hand.</p>



<p>However, recent developments have introduced a new analytic procedure that combines elements of the MLM, the SEM, and time series models called the dynamic structural equation model (or DSEM). The DSEM functionally picks up where the MLM leaves off, but expands the model to potentially include latent factors (to estimate and remove measurement error) and multiple dependent variables (to estimate reciprocal effects between two or more variables over time). DSEM is a recent development and much has yet to be learned about best practices in applied research settings, but it represents a significant development in our ability to fit complex models to ILD.</p>



<p>Want to learn more?&nbsp; We recently had the honor of being invited to provide a series of three lectures on intensive longitudinal data analysis for the American Psychological Association and we have posted our lecture materials in the resources section of the CenterStat home page (<a href="https://centerstat.org/apa-ild/" target="_blank" rel="noreferrer noopener">https://centerstat.org/apa-ild/</a>). The first session discusses the challenges and opportunities of ILD; the second focuses on the analysis of ILD using the multilevel model; and the third focuses on the analysis of ILD using the dynamic structural equation model. In addition to those resources, below are several suggested readings on the design, collection, and analysis of intensive longitudinal data. Asynchronous access to CenterStat workshops on <a href="https://centerstat.org/multilevel-modeling-async/" target="_blank" rel="noreferrer noopener"><em>Multilevel Modeling</em></a><em> </em>and <a href="https://centerstat.org/analyzing-intensive-longitudinal-data-async/" target="_blank" rel="noreferrer noopener"><em>Analyzing Intensive Longitudinal Data</em></a> is also available to those who might wish to register for additional training. You can also check our <a href="https://centerstat.org/workshops/#livestream" target="_blank" rel="noreferrer noopener">workshop schedule</a> for upcoming live offerings.</p>



<p>Good luck with your work!</p>



<p>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp; Asparouhov, T., Hamaker, E. L., &amp; Muthén, B. (2018). Dynamic structural equation models. Structural Equation Modeling: <em>A Multidisciplinary Journal, 25</em>, 359-388.</p>



<p>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp; Asparouhov, T., &amp; Muthén, B. (2020). Comparison of models for the analysis of intensive longitudinal data. <em>Structural Equation Modeling: A Multidisciplinary Journal, 27</em>, 275-297.</p>



<p>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp; Bolger, N., &amp; Laurenceau, J. P. (2013). <em>Intensive longitudinal methods: An introduction to diary and experience sampling research</em>. Guilford Press.</p>



<p>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp; Hamaker, E. L., Asparouhov, T., Brose, A., Schmiedek, F., &amp; Muthén, B. (2018). At the frontiers of modeling intensive longitudinal data: Dynamic structural equation models for the affective measurements from the COGITO study. <em>Multivariate Behavioral Research, 53</em>, 820-841.</p>



<p>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp; Hoffman, L. (2015). Longitudinal analysis: Modeling within-person fluctuation and change. Routledge.</p>



<p>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp; McNeish, D., &amp; Hamaker, E. L. (2020). A primer on two-level dynamic structural equation models for intensive longitudinal data in Mplus. <em>Psychological Methods, 25</em>, 610-635.</p>



<p>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp; McNeish, D., Mackinnon, D. P., Marsch, L. A., &amp; Poldrack, R. A. (2021). Measurement in intensive longitudinal data. <em>Structural Equation Modeling: A Multidisciplinary Journal, 28</em>, 807-822.</p>



<p>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp; Walls, T. A., &amp; Schafer, J. L. (Eds.). (2006). <em>Models for intensive longitudinal data</em>. Oxford University Press.</p>
<p>The post <a href="https://centerstat.org/ild-help-desk/">What exactly qualifies as intensive longitudinal data and why am I not able to use more traditional growth models to study stability and change over time?</a> appeared first on <a href="https://centerstat.org">CenterStat</a>.</p>
]]></content:encoded>
					
		
		
			</item>
	</channel>
</rss>

<!--
Performance optimized by W3 Total Cache. Learn more: https://www.boldgrid.com/w3-total-cache/

Object Caching 275/684 objects using Redis
Page Caching using Disk: Enhanced 

Served from: centerstat.org @ 2026-04-14 13:44:54 by W3 Total Cache
-->