@@ -769,24 +769,24 @@ \section*{Results}
769769constant, even as participants' overall level of knowledge varies across
770770quizzes or regions of the embedding space.
771771
772- Knowledge estimates need not be limited to the content of the lectures. As
773- illustrated in Figure~\ref {fig:knowledge-maps }, our general approach to
774- estimating knowledge from a small number of quiz questions may be extended to
775- \textit {any } content, given its text embedding coordinate. To visualize how
776- knowledge `` spreads'' through text embedding space to content beyond the
777- lectures participants watched, we first fit a new topic model to the lectures'
778- sliding windows with $ k = 100 $ ~topics. Conceptually, increasing the
779- number of topics used by the model functions to increase the `` resolution '' of
780- the embedding space, providing a greater ability to estimate knowledge for
781- content that is highly similar to (but not precisely the same as) that
782- contained in the two lectures. We note that we used these 2D maps solely for
783- visualization; all relevant comparisons, distance computations, and statistical
784- tests we report above were carried out in the original 15-dimensional space,
785- using the 15-topic model. Aside from increasing the number of topics from 15 to
786- 100, all other procedures and model parameters were carried over from the
787- preceding analyses. As in our other analyses, we resampled each lecture's topic
788- trajectory to 1~Hz and projected each question into a shared text embedding
789- space.
772+ Knowledge estimates need not be limited to the contents of these particular
773+ lectures and quizzes. As illustrated in Figure~\ref {fig:knowledge-maps }, our
774+ general approach to estimating knowledge from a small number of quiz questions
775+ may be extended to \textit {any } content, given its text embedding coordinate. To
776+ visualize how knowledge `` spreads'' through text embedding space to content
777+ beyond the lectures participants watched and the questions they answered, we
778+ first fit a new topic model to the lectures' sliding windows with $ k =
779+ 100 $ ~topics. Conceptually, increasing the number of topics used by the model
780+ functions to increase the `` resolution '' of the embedding space, providing a
781+ greater ability to estimate knowledge for content that is highly similar to (but
782+ not precisely the same as) that contained in the two lectures used to train the
783+ model. We note that we used these 2D maps solely for visualization; all relevant
784+ comparisons, distance computations, and statistical tests we report above were
785+ carried out in the original 15-dimensional space, using the 15-topic model.
786+ Aside from increasing the number of topics from 15 to 100, all other procedures
787+ and model parameters were carried over from the preceding analyses. As in our
788+ other analyses, we resampled each lecture's topic trajectory to 1~Hz and
789+ projected each question into a shared text embedding space.
790790
791791\begin {figure }[tp]
792792 \centering
0 commit comments