Skip to content

Commit ffd7eeb

Browse files
committed
2 parents 83d097c + 813a3c1 commit ffd7eeb

File tree

6 files changed

+111
-185
lines changed

6 files changed

+111
-185
lines changed

paper/CDL-bibliography/cdl.bib

Lines changed: 80 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -1,7 +1,80 @@
11

22

3+
@article{NortEtal02,
4+
author = {North, B V and Curtis, D and Sham, P C},
5+
date-added = {2024-02-21 18:40:44 -0500},
6+
date-modified = {2024-02-21 18:43:17 -0500},
7+
doi = {10.1086/341527},
8+
journal = {American Journal of Human Genetics},
9+
month = {Aug},
10+
number = {2},
11+
pages = {439--441},
12+
title = {A note on the calculation of empirical P values from Monte Carlo procedures.},
13+
volume = {71},
14+
year = {2002}}
15+
16+
@book{DaviHink97,
17+
author = {Davison, A. C. and Hinkley, D. V.},
18+
date-modified = {2024-02-21 18:39:11 -0500},
19+
month = {October},
20+
publisher = {Cambridge University Press},
21+
series = {Cambridge Series in Statistical and Probabilistic Mathematics},
22+
title = {Bootstrap Methods and their Application},
23+
year = {1997}}
24+
25+
@incollection{SnijBosk11,
26+
author = {T A B Snijders and R Bosker},
27+
booktitle = {{Multilevel Analysis: An Introduction to Basic and Advanced Multilevel Modeling}},
28+
chapter = {6},
29+
date-added = {2024-02-21 18:08:36 -0500},
30+
date-modified = {2024-02-21 18:29:12 -0500},
31+
edition = {2nd},
32+
month = {November},
33+
pages = {94--108},
34+
publisher = {Sage Publications},
35+
title = {More powerful tests for variance parameters},
36+
year = {2011}}
37+
38+
@article{ScheEtal08b,
39+
author = {Fabian Scheipl and Sonja Greven and Helmut K{\"u}chenhoff},
40+
date-added = {2024-02-21 17:29:35 -0500},
41+
date-modified = {2024-02-21 17:30:19 -0500},
42+
doi = {https://doi.org/10.1016/j.csda.2007.10.022},
43+
issn = {0167--9473},
44+
journal = {Computational Statistics \& Data Analysis},
45+
number = {7},
46+
pages = {3283-3299},
47+
title = {Size and power of tests for a zero random effect variance or polynomial regression in additive and linear mixed models},
48+
volume = {52},
49+
year = {2008}}
50+
51+
@article{GoldSimo00,
52+
author = {Nick Goldman and Simon Whelan},
53+
date-added = {2024-02-21 17:18:14 -0500},
54+
date-modified = {2024-02-21 17:19:25 -0500},
55+
doi = {10.1093/oxfordjournals.molbev.a026378},
56+
issn = {0737--4038},
57+
journal = {Molecular Biology and Evolution},
58+
month = {06},
59+
number = {6},
60+
pages = {975-978},
61+
title = {{Statistical Tests of Gamma-Distributed Rate Heterogeneity in Models of Sequence Evolution in Phylogenetics}},
62+
volume = {17},
63+
year = {2000}}
64+
65+
@article{HaleHojs14,
66+
author = {Halekoh, Ulrich and H{\o}jsgaard, S{\o}ren},
67+
date-added = {2024-02-21 17:05:30 -0500},
68+
date-modified = {2024-02-21 17:05:45 -0500},
69+
doi = {10.18637/jss.v059.i09},
70+
journal = {Journal of Statistical Software},
71+
number = {9},
72+
pages = {1--32},
73+
title = {{A Kenward-Roger Approximation and Parametric Bootstrap Methods for Tests in Linear Mixed Models -- The R Package pbkrtest}},
74+
volume = {59},
75+
year = {2014}}
76+
377
@article{BarrEtal13,
4-
abstract = {Linear mixed-effects models (LMEMs) have become increasingly prominent in psycholinguistics and related areas. However, many researchers do not seem to appreciate how random effects structures affect the generalizability of an analysis. Here, we argue that researchers using LMEMs for confirmatory hypothesis testing should minimally adhere to the standards that have been in place for many decades. Through theoretical arguments and Monte Carlo simulation, we show that LMEMs generalize best when they include the maximal random effects structure justified by the design. The generalization performance of LMEMs including data-driven random effects structures strongly depends upon modeling criteria and sample size, yielding reasonable results on moderately-sized samples when conservative criteria are used, but with little or no power advantage over maximal models. Finally, random-intercepts-only LMEMs used on within-subjects and/or within-items data from populations where subjects and/or items vary in their sensitivity to experimental manipulations always generalize worse than separate F1 and F2 tests, and in many cases, even worse than F1 alone. Maximal LMEMs should be the `gold standard' for confirmatory hypothesis testing in psycholinguistics and beyond.},
578
author = {Dale J. Barr and Roger Levy and Christoph Scheepers and Harry J. Tily},
679
date-added = {2024-02-20 10:18:31 -0500},
780
date-modified = {2024-02-20 10:18:52 -0500},
@@ -25798,7 +25871,8 @@ @article{GoloTaub99
2579825871
publisher = {Society for Neuroscience},
2579925872
title = {Head direction cells in rats with hippocampal or overlying neocortical lesions: evidence for impaired angular path integration},
2580025873
volume = {19},
25801-
year = {1999}}
25874+
year = {1999},
25875+
bdsk-file-1 = {YnBsaXN0MDDSAQIDBFxyZWxhdGl2ZVBhdGhZYWxpYXNEYXRhXxA1Li4vLi4vLi4vLi4vLi4vLi4vRG93bmxvYWRzL1MwMTY3OTQ3MzA3MDA0MzA2ICgxKS5iaWJPEQF0AAAAAAF0AAIAAAxNYWNpbnRvc2ggSEQAAAAAAAAAAAAAAAAAAADhKyW+QkQAAf////8ZUzAxNjc5NDczMDcwMDQzMDYgKDEpLmJpYgAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA/////+H740kAAAAAAAAAAAAGAAIAAAogY3UAAAAAAAAAAAAAAAAACURvd25sb2FkcwAAAgAyLzpVc2VyczpwYXh0b246RG93bmxvYWRzOlMwMTY3OTQ3MzA3MDA0MzA2ICgxKS5iaWIADgA0ABkAUwAwADEANgA3ADkANAA3ADMAMAA3ADAAMAA0ADMAMAA2ACAAKAAxACkALgBiAGkAYgAPABoADABNAGEAYwBpAG4AdABvAHMAaAAgAEgARAASADBVc2Vycy9wYXh0b24vRG93bmxvYWRzL1MwMTY3OTQ3MzA3MDA0MzA2ICgxKS5iaWIAEwABLwAAFQACAA3//wAAAAgADQAaACQAXAAAAAAAAAIBAAAAAAAAAAUAAAAAAAAAAAAAAAAAAAHU}}
2580225876

2580325877
@article{RobeEtal99,
2580425878
author = {R G Robertson and E T Rolls and P Georges-Fran{\c{c}}ois and S Panzeri},
@@ -31915,8 +31989,9 @@ @article{Srin99
3191531989
volume = {1},
3191631990
year = {1999}}
3191731991

31918-
@article{ScheEtal08,
31992+
@article{ScheEtal08a,
3191931993
author = {C A Schevon and S K Ng and J Cappell and R R Goodman and G McKhann and A Waziri and A Branner and A Sosunov and C E Schroeder and R G Emerson},
31994+
date-modified = {2024-02-21 17:30:06 -0500},
3192031995
journal = {Journal of Clinical Neurophysiology},
3192131996
number = {6},
3192231997
pages = {321--330},
@@ -35180,7 +35255,8 @@ @article{DawEtal02
3518035255
pages = {603--616},
3518135256
title = {Opponent interactions between serotonin and dopamine},
3518235257
volume = {15},
35183-
year = {2002}}
35258+
year = {2002},
35259+
bdsk-file-1 = {YnBsaXN0MDDSAQIDBFxyZWxhdGl2ZVBhdGhZYWxpYXNEYXRhXxArLi4vLi4vLi4vLi4vLi4vLi4vRG93bmxvYWRzL3BtY18zNzkxNzgubmJpYk8RAUwAAAAAAUwAAgAADE1hY2ludG9zaCBIRAAAAAAAAAAAAAAAAAAAAOErJb5CRAAB/////w9wbWNfMzc5MTc4Lm5iaWIAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAD/////4fv0EgAAAAAAAAAAAAYAAgAACiBjdQAAAAAAAAAAAAAAAAAJRG93bmxvYWRzAAACACgvOlVzZXJzOnBheHRvbjpEb3dubG9hZHM6cG1jXzM3OTE3OC5uYmliAA4AIAAPAHAAbQBjAF8AMwA3ADkAMQA3ADgALgBuAGIAaQBiAA8AGgAMAE0AYQBjAGkAbgB0AG8AcwBoACAASABEABIAJlVzZXJzL3BheHRvbi9Eb3dubG9hZHMvcG1jXzM3OTE3OC5uYmliABMAAS8AABUAAgAN//8AAAAIAA0AGgAkAFIAAAAAAAACAQAAAAAAAAAFAAAAAAAAAAAAAAAAAAABog==}}
3518435260

3518535261
@article{KirsEtal00,
3518635262
author = {M P Kirschen and M J Kahana and R Sekuler and B Burack},

paper/changes.pdf

5.71 KB
Binary file not shown.

paper/changes.tex

Lines changed: 19 additions & 91 deletions
Original file line numberDiff line numberDiff line change
@@ -1,7 +1,7 @@
11
\documentclass[10pt]{article}
22
%DIF LATEXDIFF DIFFERENCE FILE
33
%DIF DEL old.tex Mon Feb 19 07:49:49 2024
4-
%DIF ADD main.tex Wed Feb 21 12:35:29 2024
4+
%DIF ADD main.tex Wed Feb 21 19:15:05 2024
55
%DIF 2a2
66
\usepackage{amsmath} %DIF >
77
%DIF -------
@@ -1745,13 +1745,11 @@ \subsubsection*{Estimating dynamic knowledge traces}\label{subsec:traces}
17451745
Fundamental Forces}}\DIFadd{, }\textit{\DIFadd{Birth of Stars}}\DIFadd{, or general physics knowledge.
17461746
Note that with our coding scheme, identifiers for each }\texttt{\DIFadd{question}} \DIFadd{are
17471747
implicitly nested within levels of }\texttt{\DIFadd{lecture}} \DIFadd{and do not require explicit
1748-
nesting in our model formula.
1748+
nesting in our model formula. We then iteratively removed random effects from
1749+
the maximal model until it successfully converged with a full rank (i.e., non-singular)
1750+
random effects variance-covariance matrix.
17491751
}
17501752
1751-
%DIF > We then iteratively removed random effects from the maximal model until it
1752-
%DIF > successfully converged with a full rank (i.e., non-singular) random effects
1753-
%DIF > variance-covariance matrix.
1754-
17551753
%DIF > % JRM NOTE: do we need this next paragraph? Commenting out for now...
17561754
%DIF > %When inspecting the model's random effect estimates revealed multiple terms estimated at the boundary of their parameter space (i.e., variance components of 0 or correlation terms of $\pm 1$), we found that the order in which we eliminated these terms typically did not affect which terms did and did not need to be removed in order for the model to converge to a non-degenerate solution.
17571755
%DIF > When this required eliminating multiple terms whose estimates reached the boundary of their parameter space (i.e., variance components of 0 or correlation terms of $\pm 1$), we found that the order in which we did so typically did not change the set of terms that needed to be removed in order for the model to converge to a non-degenerate solution.
@@ -1775,98 +1773,28 @@ \subsubsection*{Estimating dynamic knowledge traces}\label{subsec:traces}
17751773
17761774
\DIFadd{To assess the predictive value of our knowledge estimates, we compared each
17771775
GLMM's ability to discriminate between correctly and incorrectly answered
1778-
questions to that of an analogous model that did not consider estimated
1776+
questions to that of an analogous model that did }\textit{\DIFadd{not}} \DIFadd{consider estimated
17791777
knowledge. Specifically, we used the same sets of observations with which we
17801778
fit each ``full'' model to fit a second ``null'' model, with the formula:
17811779
}\[
17821780
\DIFadd{\mathtt{accuracy \sim (1\ \vert\ participant) + (1\ \vert\ question)}
17831781
}\]
1784-
\DIFadd{where the terms are as defined above.
1782+
\DIFadd{where ``}\texttt{\DIFadd{accuracy}}\DIFadd{'', ``}\texttt{\DIFadd{participant}}\DIFadd{'', and ``}\texttt{\DIFadd{question}}\DIFadd{'' are as defined above.
1783+
As with our full models, the null models we fit for the “All questions” version of the analysis for each quiz contained an additional term, $\mathtt{(1\ \vert\ lecture)}$, where ``}\texttt{\DIFadd{lecture}}\DIFadd{'' are as defined above.
1784+
We then compared each full model to its reduced (null) equivalent using a likelihood-ratio test (LRT).
1785+
Because the typical asymptotic $\chi^2_d$ approximation of the null distribution for the LRT statistic ($\lambda_{LR}$) is anti-conservative for models that differ in their random slope terms~\mbox{%DIFAUXCMD
1786+
\citep{GoldSimo00,ScheEtal08b,SnijBosk11}}\hskip0pt%DIFAUXCMD
1787+
, we computed $p$-values for these tests using a parametric bootstrapping procedure~\mbox{%DIFAUXCMD
1788+
\citep{HaleHojs14}}\hskip0pt%DIFAUXCMD
1789+
.
1790+
For each of 1,000 bootstraps, we used the fitted null model to simulate a sample of observations of equal size to our original sample.
1791+
We then re-fit both the null and full models to this simulated sample and compared them via an LRT.
1792+
This yielded a distribution of $\lambda_{LR}$ statistics we may expect to observe under our null hypothesis.
1793+
Following~\mbox{%DIFAUXCMD
1794+
\citep{DaviHink97,NortEtal02}}\hskip0pt%DIFAUXCMD
1795+
, we computed a corrected $p$-value for our observed $\lambda_{LR}$ as $\frac{r + 1}{n + 1}$, where $r$ is the number of simulated model comparisons that yielded a $\lambda_{LR}$ greater than or equal to our observed value and $n$ is the number of simulations we ran (1,000).
17851796
}
17861797
1787-
1788-
1789-
1790-
1791-
1792-
1793-
1794-
1795-
%DIF > In order to assess the predictive value of the knowledge estimates, we then fit a second set of ``null'' models to the same sets of observations used to fit our full GLMMs. These
1796-
%DIF >
1797-
%DIF >
1798-
%DIF > used the same sets of observations used to fit these GLMMs to fit a second set of ``null'' models.
1799-
%DIF >
1800-
%DIF >
1801-
%DIF >
1802-
%DIF > Next, in order to assess the predictive value of the knowledge estimates
1803-
%DIF >
1804-
%DIF >
1805-
%DIF >
1806-
%DIF > In order to assess the predictive value of the knowledge estimates we used to fit each GLMM, we then fit a \textit{second} model to the same data as each of the 15
1807-
%DIF >
1808-
%DIF > In order to assess whether the knowledge estimates we used to fit each GLMM could reliably predict participants' success on held-out questions, we then fit a second GLMM to the observations
1809-
1810-
1811-
%DIF > $$
1812-
%DIF > \begin{aligned}
1813-
%DIF > \operatorname{accuracy}_{i} &\sim \operatorname{Binomial}(n = 1, \operatorname{prob}_{\operatorname{accuracy} = \operatorname{correct}} = \widehat{P}) \\
1814-
%DIF > \log\left[\frac{\hat{P}}{1 - \hat{P}} \right] &=\alpha_{j[i],k[i]} \\
1815-
%DIF > \left(
1816-
%DIF > \begin{array}{c}
1817-
%DIF > \begin{aligned}
1818-
%DIF > &\alpha_{j} \\
1819-
%DIF > &\gamma_{1j}
1820-
%DIF > \end{aligned}
1821-
%DIF > \end{array}
1822-
%DIF > \right)
1823-
%DIF > &\sim N \left(
1824-
%DIF > \left(
1825-
%DIF > \begin{array}{c}
1826-
%DIF > \begin{aligned}
1827-
%DIF > &\gamma_{0}^{\alpha} + \gamma_{1k[i]}^{\alpha}(\operatorname{knowledge}) \\
1828-
%DIF > &\mu_{\gamma_{1j}}
1829-
%DIF > \end{aligned}
1830-
%DIF > \end{array}
1831-
%DIF > \right)
1832-
%DIF > ,
1833-
%DIF > \left(
1834-
%DIF > \begin{array}{cc}
1835-
%DIF > \sigma^2_{\alpha_{j}} & \rho_{\alpha_{j}\gamma_{1j}} \\
1836-
%DIF > \rho_{\gamma_{1j}\alpha_{j}} & \sigma^2_{\gamma_{1j}}
1837-
%DIF > \end{array}
1838-
%DIF > \right)
1839-
%DIF > \right)
1840-
%DIF > \text{, for participant j = 1,} \dots \text{,J} \\
1841-
%DIF > \left(
1842-
%DIF > \begin{array}{c}
1843-
%DIF > \begin{aligned}
1844-
%DIF > &\alpha_{k} \\
1845-
%DIF > &\gamma_{1k}
1846-
%DIF > \end{aligned}
1847-
%DIF > \end{array}
1848-
%DIF > \right)
1849-
%DIF > &\sim N \left(
1850-
%DIF > \left(
1851-
%DIF > \begin{array}{c}
1852-
%DIF > \begin{aligned}
1853-
%DIF > &\mu_{\alpha_{k}} \\
1854-
%DIF > &\mu_{\gamma_{1k}}
1855-
%DIF > \end{aligned}
1856-
%DIF > \end{array}
1857-
%DIF > \right)
1858-
%DIF > ,
1859-
%DIF > \left(
1860-
%DIF > \begin{array}{cc}
1861-
%DIF > \sigma^2_{\alpha_{k}} & \rho_{\alpha_{k}\gamma_{1k}} \\
1862-
%DIF > \rho_{\gamma_{1k}\alpha_{k}} & \sigma^2_{\gamma_{1k}}
1863-
%DIF > \end{array}
1864-
%DIF > \right)
1865-
%DIF > \right)
1866-
%DIF > \text{, for question k = 1,} \dots \text{,K}
1867-
%DIF > \end{aligned}
1868-
%DIF > $$
1869-
18701798
\DIFaddend \subsubsection*{Estimating the ``smoothness'' of knowledge}\label{subsec:smoothness}
18711799
18721800
In the analysis reported in Figure~\ref{fig:smoothness}A, we show how

paper/main.pdf

4.26 KB
Binary file not shown.

paper/main.tex

Lines changed: 12 additions & 90 deletions
Original file line numberDiff line numberDiff line change
@@ -1502,11 +1502,9 @@ \subsubsection*{Generalized linear mixed models}\label{subsec:glmm}
15021502
Fundamental Forces}, \textit{Birth of Stars}, or general physics knowledge.
15031503
Note that with our coding scheme, identifiers for each \texttt{question} are
15041504
implicitly nested within levels of \texttt{lecture} and do not require explicit
1505-
nesting in our model formula.
1506-
1507-
% We then iteratively removed random effects from the maximal model until it
1508-
% successfully converged with a full rank (i.e., non-singular) random effects
1509-
% variance-covariance matrix.
1505+
nesting in our model formula. We then iteratively removed random effects from
1506+
the maximal model until it successfully converged with a full rank (i.e., non-singular)
1507+
random effects variance-covariance matrix.
15101508

15111509
%% JRM NOTE: do we need this next paragraph? Commenting out for now...
15121510
% %When inspecting the model's random effect estimates revealed multiple terms estimated at the boundary of their parameter space (i.e., variance components of 0 or correlation terms of $\pm 1$), we found that the order in which we eliminated these terms typically did not affect which terms did and did not need to be removed in order for the model to converge to a non-degenerate solution.
@@ -1531,96 +1529,20 @@ \subsubsection*{Generalized linear mixed models}\label{subsec:glmm}
15311529

15321530
To assess the predictive value of our knowledge estimates, we compared each
15331531
GLMM's ability to discriminate between correctly and incorrectly answered
1534-
questions to that of an analogous model that did not consider estimated
1532+
questions to that of an analogous model that did \textit{not} consider estimated
15351533
knowledge. Specifically, we used the same sets of observations with which we
15361534
fit each ``full'' model to fit a second ``null'' model, with the formula:
15371535
\[
15381536
\mathtt{accuracy \sim (1\ \vert\ participant) + (1\ \vert\ question)}
15391537
\]
1540-
where the terms are as defined above.
1541-
1542-
1543-
1544-
1545-
1546-
1547-
1548-
1549-
1550-
%In order to assess the predictive value of the knowledge estimates, we then fit a second set of ``null'' models to the same sets of observations used to fit our full GLMMs. These
1551-
%
1552-
%
1553-
%used the same sets of observations used to fit these GLMMs to fit a second set of ``null'' models.
1554-
%
1555-
%
1556-
%
1557-
%Next, in order to assess the predictive value of the knowledge estimates
1558-
%
1559-
%
1560-
%
1561-
%In order to assess the predictive value of the knowledge estimates we used to fit each GLMM, we then fit a \textit{second} model to the same data as each of the 15
1562-
%
1563-
%In order to assess whether the knowledge estimates we used to fit each GLMM could reliably predict participants' success on held-out questions, we then fit a second GLMM to the observations
1564-
1565-
1566-
%$$
1567-
%\begin{aligned}
1568-
% \operatorname{accuracy}_{i} &\sim \operatorname{Binomial}(n = 1, \operatorname{prob}_{\operatorname{accuracy} = \operatorname{correct}} = \widehat{P}) \\
1569-
% \log\left[\frac{\hat{P}}{1 - \hat{P}} \right] &=\alpha_{j[i],k[i]} \\
1570-
%\left(
1571-
% \begin{array}{c}
1572-
% \begin{aligned}
1573-
% &\alpha_{j} \\
1574-
% &\gamma_{1j}
1575-
% \end{aligned}
1576-
% \end{array}
1577-
%\right)
1578-
% &\sim N \left(
1579-
%\left(
1580-
% \begin{array}{c}
1581-
% \begin{aligned}
1582-
% &\gamma_{0}^{\alpha} + \gamma_{1k[i]}^{\alpha}(\operatorname{knowledge}) \\
1583-
% &\mu_{\gamma_{1j}}
1584-
% \end{aligned}
1585-
% \end{array}
1586-
%\right)
1587-
%,
1588-
%\left(
1589-
% \begin{array}{cc}
1590-
% \sigma^2_{\alpha_{j}} & \rho_{\alpha_{j}\gamma_{1j}} \\
1591-
% \rho_{\gamma_{1j}\alpha_{j}} & \sigma^2_{\gamma_{1j}}
1592-
% \end{array}
1593-
%\right)
1594-
% \right)
1595-
% \text{, for participant j = 1,} \dots \text{,J} \\
1596-
%\left(
1597-
% \begin{array}{c}
1598-
% \begin{aligned}
1599-
% &\alpha_{k} \\
1600-
% &\gamma_{1k}
1601-
% \end{aligned}
1602-
% \end{array}
1603-
%\right)
1604-
% &\sim N \left(
1605-
%\left(
1606-
% \begin{array}{c}
1607-
% \begin{aligned}
1608-
% &\mu_{\alpha_{k}} \\
1609-
% &\mu_{\gamma_{1k}}
1610-
% \end{aligned}
1611-
% \end{array}
1612-
%\right)
1613-
%,
1614-
%\left(
1615-
% \begin{array}{cc}
1616-
% \sigma^2_{\alpha_{k}} & \rho_{\alpha_{k}\gamma_{1k}} \\
1617-
% \rho_{\gamma_{1k}\alpha_{k}} & \sigma^2_{\gamma_{1k}}
1618-
% \end{array}
1619-
%\right)
1620-
% \right)
1621-
% \text{, for question k = 1,} \dots \text{,K}
1622-
%\end{aligned}
1623-
%$$
1538+
where ``\texttt{accuracy}'', ``\texttt{participant}'', and ``\texttt{question}'' are as defined above.
1539+
As with our full models, the null models we fit for the “All questions” version of the analysis for each quiz contained an additional term, $\mathtt{(1\ \vert\ lecture)}$, where ``\texttt{lecture}'' are as defined above.
1540+
We then compared each full model to its reduced (null) equivalent using a likelihood-ratio test (LRT).
1541+
Because the typical asymptotic $\chi^2_d$ approximation of the null distribution for the LRT statistic ($\lambda_{LR}$) is anti-conservative for models that differ in their random slope terms~\citep{GoldSimo00,ScheEtal08b,SnijBosk11}, we computed $p$-values for these tests using a parametric bootstrapping procedure~\citep{HaleHojs14}.
1542+
For each of 1,000 bootstraps, we used the fitted null model to simulate a sample of observations of equal size to our original sample.
1543+
We then re-fit both the null and full models to this simulated sample and compared them via an LRT.
1544+
This yielded a distribution of $\lambda_{LR}$ statistics we may expect to observe under our null hypothesis.
1545+
Following~\citep{DaviHink97,NortEtal02}, we computed a corrected $p$-value for our observed $\lambda_{LR}$ as $\frac{r + 1}{n + 1}$, where $r$ is the number of simulated model comparisons that yielded a $\lambda_{LR}$ greater than or equal to our observed value and $n$ is the number of simulations we ran (1,000).
16241546

16251547
\subsubsection*{Estimating the ``smoothness'' of knowledge}\label{subsec:smoothness}
16261548

paper/supplement.pdf

0 Bytes
Binary file not shown.

0 commit comments

Comments
 (0)