How do people perceive and communicate structure? We investigate this question by letting participants play a communication game, where one player describes a pattern, and another player redraws it based on the description alone. We use this paradigm to compare two models of pattern description, one compositional (complex structures built out of simpler ones) and one noncompositional. We find that compositional patterns are communicated more effectively than noncompositional patterns, that a compositional model of pattern description predicts which patterns are harder to describe, and that this model can be used to evaluate participants’ drawings, producing humanlike quality ratings. Our results suggest that natural language can tap into a compositionally structured pattern description language.
Humans see patterns everywhere, and eagerly communicate them to one another. However, little is known formally about how we communicate patterns, what kinds of patterns are easier or harder to communicate, and how we reconstruct patterns from natural language. This article seeks to bridge this gap by combining a pattern communication game with a mathematical model of pattern description (Quiroga, Schulz, Speekenbrink, & Harvey, 2018; Schulz, Tenenbaum, Duvenaud, Speekenbrink, & Gershman, 2017).
Consider the graphs shown in Figure 1, which plot time series of CO2 emission, airline passenger volume, and search frequency for the term “gym membership.” Experiments suggest that humans perceive these graphs as compositions of simpler patterns, such as lines, oscillations, and smoothly changing curves (Quiroga et al., 2018; Schulz, Tenenbaum, et al., 2017). For example, there is seasonal variation in passenger volume (a periodic component with time-dependent amplitude), superimposed on a linear increase over time.
As described in more detail in the next section, we can formalize this idea using a pattern description language consisting of functional primitives and algebraic operations that compose them together. By defining a probability distribution over this description language, we can express an inductive bias for certain kinds of functions—in particular, functions that can be described with a small number of compositions (Duvenaud, Lloyd, Grosse, Tenenbaum, & Ghahramani, 2013; Lloyd, Duvenaud, Grosse, Tenenbaum, & Ghahramani, 2014; Schulz, Tenenbaum, et al., 2017). In other words, the “mental” description length of a function relates to the complexity of its encoding in the compositional pattern description language.
It is important to note that there are other ways to reduce description length besides encoding functions with a small set of compositions (what we will refer to as “compositional functions”). For example, a standard assumption in machine learning is that functions are smooth (Rasmussen & Williams, 2006). If we defined a probability distribution over functions that prefer smoothness, then smooth functions would have short description lengths, in the sense that the number of bits required to encode them would be smaller than nonsmooth functions. However, a preference for smoothness does not seem to be an adequate account of how humans encode functions: functions that are smooth but cannot be compactly described by compositions are less easily encoded, as indicated by poorer memory and change detection performance for these functions compared to compositional functions (Schulz, Tenenbaum, et al., 2017; see additional analysis in the Supplemental Materials).
Here we extend this idea one step further, asking whether there is a correspondence between the pattern description language and natural language descriptions of functions. We proceed in three steps. First, we ask participants to describe functions sampled from compositional or noncompositional distributions. Second, we ask a separate group of participants to redraw the original function using only the description. Third, we ask another group of participants to rate how well each drawing corresponds to the original. We hypothesized that compositional functions would be easier to reconstruct compared to noncompositional functions, under the assumption that the former allow for a mental description that can be more easily encoded into natural language and decoded back into the function space. We also rule out several alternative explanations and map pattern-specific descriptions to compositional components with the help of an additional experiment.
A COMPOSITIONAL PATTERN DESCRIPTION LANGUAGE
All positive semidefinite kernels are closed under addition and multiplication, allowing us to create richly structured and interpretable kernels from well-understood base components. We use this property to construct a class of compositional kernels (Duvenaud et al., 2013; Lloyd et al., 2014; Schulz, Tenenbaum, et al., 2017). To give some intuition for this approach, consider again the CO2 data in Figure 1. This function is naturally decomposed into a sum of a linearly increasing component and a seasonally periodic component. The compositional kernel captures this structure by summing a linear and periodic kernel.
Compositional GPs have been used to model complex time-series data (Duvenaud et al., 2013), as well as to generate automated natural language descriptions from data (Lloyd et al., 2014), an approach coined the “automated statistician” (Ghahramani, 2015). Although it is frequently assumed that people will easily understand the generated description of the “automated statistician,” it is not known whether compositional patterns are indeed more communicable.
We follow the approach developed in Schulz, Tenenbaum, et al. (2017), using three base kernels that define basic structural patterns: a linear kernel that can encode trends, a radial basis function kernel that can encode smooth functions, and a periodic kernel that can encode repeated patterns (see Table 1). These kernels can be combined by either multiplying or adding them together. In previous research, we found that this compositional grammar can account for participants’ behavior across a variety of experimental paradigms, including pattern completions, change detection, and working, memory tasks (Schulz, Tenenbaum, et al., 2017). We fix the maximum number of combined kernels to be three and do not allow for repetition of kernels in order to restrict the complexity of inference (see next section).
|Name .||Definition .|
|Linear||k(x, x′) = (x − θ1)(x′ − θ1)|
|Radial basis||k(x, x′) =|
|Periodic||k(x, x′) =|
|Name .||Definition .|
|Linear||k(x, x′) = (x − θ1)(x′ − θ1)|
|Radial basis||k(x, x′) =|
|Periodic||k(x, x′) =|
We compare the compositional model to a noncompositional model based on a spectral mixture of kernels (see Supplemental Materials, for further details). This model is derived from the fact that any stationary kernel can be expressed as an integral using Bochner’s theorem. This model approximates functions by matching their spectral density with a mixture of Gaussians. It has a similar expressivity compared to the compositional model, but does not encode compositional structure explicitly. This means that both models will make similar predictions given unlimited data; however, given a finite data regime, the compositional kernel will have strong inductive biases for compositional functions, whereas the spectral kernel will not show such inductive biases. Wilson, Dann, Lucas, and Xing (2015) have used this model to reverse-engineer “human kernels” in standard function-learning tasks. We use this kernel to assess if communication of patterns can be described well by a kernel that is equally expressive as the compositional kernel but does not operate over structural building blocks. Instead of optimizing its parameters to find humanlike kernels in traditional function-learning tasks, we will optimize it based on the structure participants had to describe.1
MODELING FUNCTION LEARNING
We use the same patterns as in Schulz, Tenenbaum, et al. (2017). These patterns were generated from both compositional and noncompositional (spectral mixture) kernels. The compositional patterns were sampled randomly from a compositional grammar by first randomly sampling a kernel composition and then sampling a function from that kernel, whereas the noncompositional patterns were sampled from the spectral mixture kernel, where the number of components was varied between two and six uniformly. A subset of these sampled patterns were then chosen so that compositional and noncompositional functions were matched based on their spectral entropy and wavelet distance (Goerg, 2013), leading to a final set of 40 patterns.
PATTERN COMMUNICATION GAME
Our study assessed how well different patterns can be communicated in a free-form communication game (i.e., without any restrictions on participants’ description lengths or word usage). The study consisted of three parts: description, drawing, and quality rating. Participants were recruited from Amazon Mechanical Turk, and no participant was allowed to participate in more than one part. The study was approved by Harvard’s institutional review board.
Part 1: Eliciting Descriptions
Thirty-one participants (6 female, mean age = 34.91, SD = 10.25) took part in the description study. Participants sequentially saw six different patterns, represented as graphs that they had to describe afterwards. Three of the patterns were randomly sampled from the 20 compositional patterns without replacement, and three were sampled from the noncompositional pool of patterns. The order of the presented patterns was determined at random. On every trial, participants first saw a pattern for 10 s, after which the pattern disappeared. The pattern was shown to them as 100 equidistant points indicating a function on a canvas (see Figure 2). After the pattern disappeared, participants had to describe it using as many words as they liked. Participants were told that we would pass on their descriptions to someone else who would then have to redraw the patterns without ever having seen them.
Two judges independently rated the descriptions2 on a scale from 1 (bad descriptions) to 5 (great descriptions). The agreement between the two judges was sufficiently high, with an interrater correlation of r(29) = 0.46, t = 2.45, p = .02, BF = 3.8, and we validated their judgments both statistically and using additional raters (see the Supplemental Materials). We then retained the descriptions with an average rating higher than 3, giving 7 “describers” and a total pool of 31 different patterns. Sixteen of these patterns were compositional, and fifteen were noncompositional. All participants were paid $2 for their participation.
Part 2: Drawing the Patterns
We recruited 49 participants (21 females, mean age = 33.6, SD = 9.6) for the drawing part of the experiment. In this part, participants only saw the descriptions of the patterns and had to redraw them by placing dots on an empty canvas. Below the canvas, participants saw the descriptions of the patterns, which they knew had been written by a past participant. Participants were told that they could place any number of dots onto the canvas, but had to place at least five dots to draw a pattern before they could submit their drawings. Each participant received the six descriptions written by a randomly matched participant from the description part, that is, they were paired with one of the top seven “describers” from the first part of the study. Participants were paid $2 for their participation.
Part 3: Rating the Quality of the Drawings
We recruited 104 participants (35 females, mean age = 37.7, SD = 8.6) to rate the quality of participants’ performance in the previous parts. Participants were told the rules of the game the previous participants had played. They then had to rate 30 randomly sampled drawings, where the drawings were always presented right next to the original pattern. Participants did not see the descriptions that led to the eventual drawings, but rather only had to evaluate how much the drawing resembled the original, that is, how well they thought two participants performed in one round of the game. They did this by entering values on a slider from 0 (bad performance) to 100 (great performance). We paid participants $1 for their participation.
Figure 2 shows three examples of participants’ descriptions and drawings for both compositional and noncompositional patterns. We first assessed whether participants in the description part of the study entered longer descriptions for the compositional than the noncompositional patterns. This analysis revealed no significant difference between the two kinds of patterns, t(30) = 0.15, p = .88, d = 0.03, BF = 0.2. Next, we assessed whether participants in the drawing part of the study used more dots to redraw compositional than noncompositional patterns. This also showed no difference between the two kinds of patterns, t(49) = 1.00, p = .32, d = 0.14, BF = 0.2.
Although one might conclude from these analyses that the descriptions and redrawings were relatively similar across the two pattern classes, inspection of which words frequently appeared in the compositional descriptions but not the noncompositional ones (and vice versa) revealed that compositional descriptions often included more abstract words such as “mountain,” “repeat,” or “valley” (Figure 3a), whereas noncompositional descriptions used words such as “starts,” “bottom,” or “top,” likely describing exactly how to draw a particular shape (Figure 3b). Furthermore, we assessed the descriptions’ lexical diversity, defined as the sum of the unique words used divided by all words used in a description (McCarthy & Jarvis, 2010). Compositional descriptions showed a higher lexical diversity than noncompositional descriptions, t(30) = 4.22, p < .001, d = 0.76, BF > 100, Figure 3c.
We next analyzed the quality of participants’ drawings. In order to compare the two, we used polynomial smoothing splines to connect the dots. The splines were forced to go through every point on the canvas such that the original and redrawn patterns have the same length. Our results also hold even if we just use the raw points or other methods of extracting the patterns such as generalized additive models (see the Supplemental Materials). We then calculated the absolute difference (absolute error) between the original and the redrawn patterns. This difference was larger for noncompositional than for compositional patterns (Figure 4a; t(49) = 2.43, p = .01, d = 0.34, BF = 4.1), indicating that participants were more accurate at redrawing compositional patterns.
The absolute distance between two patterns might not be the best indicator of performance, because two patterns can look alike but still show a large absolute difference (e.g., if the redrawn pattern is smaller than the original, or if one pattern is just slightly shifted to either side). We therefore also applied a distance measure that takes into account these possible deviations by assessing the similarity of two patterns based on their differences after performing a Haar wavelet transform. The idea behind this similarity measure is to replace the original pattern by its wavelet approximation coefficients, and then to measure similarity between these coefficients (Montero & Vilar, 2014; see the Supplemental Materials). Technicalities aside, this measure is robust to scaling and shifting of the patterns. We have previously verified that it corresponds well with participants’ similarity judgments when comparing two patterns (Schulz, Tenenbaum, et al., 2017). Analyzing participants’ performance using this measurement (wavelet distance) showed an even stronger advantage for compositional patterns (Figure 4b; t(49) = 3.02, p = .004, d = 0.43, BF = 11.7).
Next, we looked at the quality ratings collected in the third part of our study. We estimated a linear mixed-effects model with random effects for a compositional vs. noncompositional contrast for raters, describer-drawer pairs, and for the items (patterns). We compared this model to another model that also included a compositional vs. noncompositional contrast as a fixed effect (following the logic of Barr, Levy, Scheepers, & Tily, 2013). The results of this analysis showed that adding the compositional contrast as a fixed effect moderately improved the overall model fit (BF = 4.6). Compositional patterns were rated more highly than noncompositional patterns (Figure 4c), resulting in a posterior estimate of 39.61 (95% HDI [high density interval]: 39.03, 40.19) for the compositional patterns and a posterior estimate of 33.31 (95% HDI: 32.69, 33.93). Interestingly, the rated quality was not influenced by the length of the descriptions (BF = 0.01).
We also assessed how well both models captured the difficulty of communicating the different patterns, as well as participants’ quality ratings. First, we assessed whether the likelihood of each model, when fitted to the original patterns, was predictive of how communicable that pattern was. The idea behind this analysis was that, if participants were really using one of the two models to extract and compress patterns, then how well this model can compress the patterns (as measured by the likelihood given the data) should be related to how well people can communicate it. We therefore fitted a set of multilevel regression models with the previously used error measures as the dependent variables, and the log-likelihood for each pattern as estimated by both compositional and noncompositional models as the independent variables. We also included a random intercept and a random slope for each of the two models’ likelihoods, as participants might vary in their ability to redraw the described pattern and how well they are predicted by the different models. The resulting fixed effects regression coefficients (Table 2) showed the same pattern for both error measurements: there was a significant effect for the compositional but not the noncompositional log-likelihoods. Moreover, we directly compared two mixed-effects regressions solely using either the compositional or the noncompositional log-likelihoods as the independent variable. This comparison strongly favored the compositional log-likelihoods for modeling both the absolute error (BF > 100) and the wavelet distance (BF > 100). This means that patterns that were easier to compress by the compositional model were also easier to communicate for participants. This was not true for the noncompositional model.
|.||Absolute Error .||Wavelet Distance .||Quality Ratings .|
|Intercept||27.59** (0.63)||3.26** (0.07)||35.83** (2.38)|
|Compositional||−1.39* (0.54)||−0.19** (0.06)||6.73** (2.12)|
|Noncompositional||− 0.83 (0.53)||− 0.08 (0.06)||− 4.03 (3.15)|
|.||Absolute Error .||Wavelet Distance .||Quality Ratings .|
|Intercept||27.59** (0.63)||3.26** (0.07)||35.83** (2.38)|
|Compositional||−1.39* (0.54)||−0.19** (0.06)||6.73** (2.12)|
|Noncompositional||− 0.83 (0.53)||− 0.08 (0.06)||− 4.03 (3.15)|
Note. Columns show the standardized fixed effects regression estimates for modeling the absolute error, the wavelet distance error, or participants’ quality ratings as the dependent variable. Standard errors of the coefficients are displayed below each coefficient in brackets.
**p < .001, *p < .01
Finally, we applied the same regression approach, using the log-likelihood as the independent variable (both as a fixed and a random effect), to predict the quality ratings collected in the third part of the study. The idea behind this analysis is that if participants were indeed using one of the two models to evaluate the quality of the drawings, then they should evaluate the likelihood of the drawing to have been produced by the same generative process as the original drawing. Only the compositional model significantly predicted participant’s ratings in part 3 (Table 2 and Figure 4c) and the direct comparison between the compositional and the noncompositional model strongly favored the compositional model (BF > 100). This suggests that participants assessed the quality of the drawings based on how well they could be described by similar compositions as the original patterns.
Controlling for Individual Components
Given that both the compositional and the noncompositional kernel can—in the limit of infinite data—capture any function but, differ in their inductive biases given finite data, we also analyzed if any individual structure (for example, periodicity or linearity) might have driven the differences between compositional and noncompositional patterns’ communicability. We therefore analyzed the differences between compositional and noncompositional patterns’ wavelet distances while controlling for how well different single-component kernels described the patterns, as measured by the log-likelihoods produced by either a periodic, a linear, or an RBF kernel taken on their own. We regressed the individual components’ log-likelihoods as a fixed and a random effect onto the wavelet distances first. Additionally, we added a dummy indicating whether or not a pattern was compositional to that regression as a random effect. Afterward, we added the same dummy variable as a fixed effect to assess if compositionality added something to communicability over and above the simple components. This analysis showed that adding the dummy factor improved a regression that only contained the periodic (BF = 20.7), the RBF (BF = 28.9), or the linear (BF = 15.6) log-likelihoods. Thus, the advantage of compositional patterns’ communicability did not solely arise from single structures, persisting even when controlling for each of the individual components of the compositional grammar.
Controlling for Pattern Memorability
One concern with our current analysis is that participants saw the patterns and then had to describe them from memory. Thus, differences in the final quality could have also arisen from differences in participants’ memory capacity for different patterns. To rule out this alternative explanation, we also assessed by how much, if at all, the compositional model’s predictions captured communication quality better than just pattern memorability. We therefore ran an additional experiment in which 51 participants (37 male, mean age = 31.91, SD = 11.8) sequentially saw patterns for 10 s (just like in part 1 of our main experiment) and then had to immediately redraw it (using the same canvas setup as in part 2 of our main experiment). We let participants do this for six patterns in total. Three of these patterns were compositional and three were noncompositional. We then measured how well the different patterns could be remembered by calculating the wavelet differences between the original and the redrawn patterns and averaging them for each pattern individually, leading to an item-specific measure of memorability. Next, we assessed by how much our previous regressions improved by additionally entering the compositional model’s log-likelihoods as a fixed effect while controlling for the item-specific memorability score (both as a random and fixed effect) and the compositional model’s log-likelihoods as random effect. This revealed that the compositional model’s likelihood substantially improved the regression model for the absolute error (BF = 8.9), the wavelet distance measure (BF = 8.3), and the quality ratings (BF > 100). Thus, there are strong reasons to believe that the differences in communication qualities did not solely arise from pattern memorability.
Relating Composition-Specific Words to Compositional Descriptions
We were also interested in how specific features of participant language mapped onto specific compositions in the patterns. We therefore conducted another experiment in which we showed an additional group of participants single components of the compositional model. In this experiment, 50 participants (24 males, mean age = 34.25, SD = 11.96) saw six different patterns sequentially. Each pattern was presented to them for 10 s after which it disappeared and they had to describe it, exactly as in part 1 of our earlier experiments. However, this time we sampled patterns from single kernels of the compositional model. Thus, each participant had to describe two patterns that were sampled from a periodic kernel, two patterns sampled from an RBF kernel, and two patterns sampled from a linear kernel, presented to them in random order. We then extracted the top 10 words for each single component, that is, the words that were more frequently used to describe patterns from a particular component compared to the other two components. The resulting words were intuitively plausible; for example, common words for periodic patterns were “peak,” “time,” and “wave,” whereas frequent words for linear patterns were “linear,” “straight,” and “steady.” We then assessed how often the extracted, composition-specific words appeared in the descriptions elicited in part 1 of our earlier experiment. Figure 5a shows how much more often the extracted words appeared in the descriptions of compositional as compared to noncompositional descriptions in our first experiment (calculated by subtracting the frequency of occurrences in the noncompositional descriptions from the frequency of occurrences in the compositional descriptions). This revealed that many of the compositional words appeared more frequently in the descriptions of compositional patterns than in the descriptions of noncompositional patterns. This can also be seen when calculating—for each set of words—the probability that at least one of the words appeared in the description (Figure 5b). This probability was higher for compositional patterns overall, t(30) = 2.65, p = .005, d = 0.54, BF = 7.47. Moreover, both words describing periodic, t(30) = 4.14, p < .001, d = 0.74, BF > 100, and linear, t(30) = 3.92, p < .001, d = 0.70, BF = 63.3, patterns were more frequently used to described compositional than noncompositional patterns. This difference was not present for words describing RBF patterns, t(30) = −0.96, p = .34, d = 0.17, BF = 0.3. This is intuitive because noncompositional patterns might also contain smooth parts. Indeed, the compositional model more frequently interprets patterns sampled from the noncompositional kernel as having RBF components than linear or periodic components (cf. Schulz, Tenenbaum, et al., 2017).
Finally, we calculated for each component the probability of being present in each of the described functions. This can be approximated by dividing the summed log-likelihood of kernels containing a particular component by the sum of all log-likelihoods. We then regressed the resulting values onto a binary variable that indicated whether or not a composition-specific description was present for each description, including a random intercept over participants.3 For example, one would expect that participants might be more likely to use RBF-specific words the more likely it actually was that an RBF component was part of the seen pattern. This showed that linear words were somewhat more likely to be used the more likely linear patterns were to be present in the data, β = 0.13, z = 2.71, p = .007, BF = 3.8, 95% HDI: 0.04, 0.22, and that the same was also true for RBF-specific, β = 0.13, z = 2.64, p = .008, BF = 5.2, 95% HDI: 0.01, 0.26, and periodic-specific words, β = 0.12, z = 2.32, p = 0.02, BF = 4.1, 95% HDI: 0.02, 0.23.
We investigated how people perceive and communicate patterns in a pattern communication game where one participant described a pattern and another participant used this description to redraw the pattern. Our results provide evidence that compositional patterns are more communicable, that a compositional model better captures participants’ difficulty in communicating patterns, and that participants’ quality ratings when evaluating the performance of other participants are also best captured by a compositional model. Taken together, these results suggest that there is an interface between natural language and the compositional pattern description language uncovered by our earlier work (Schulz, Tenenbaum, et al., 2017).
We are not the first to study how patterns are transmitted from one person to another. Kalish, Griffiths, and Lewandowsky (2007) let participants learn and reproduce functional patterns in an “iterated learning” paradigm. In this paradigm, participants drew functions that were then passed on to the next person, who then had to redraw them, and so forth. The results of this study showed that participants converged to linear functions with a positive slope, even if they started out from linear functions with a negative slope or just random dots. A key difference from our study is that Kalish et al. (2007) did not ask participants to generate natural language descriptions. Another difference is that in iterated-learning studies, the object of interest is typically the stationary distribution, which reveals the learner’s inductive biases (Griffiths & Kalish, 2007; Kirby & Hurford, 2002). We have not attempted to simulate a Markov chain to convergence, so our study does not say anything about the stationary distribution. Here we ask whether particular pattern classes are more or less communicable. Schulz, Tenenbaum, et al. (2017) provide a systematic investigation into the nature of inductive biases in function learning, supporting the claim that these inductive biases are compositional in nature.
Our approach ties together neatly with past attempts to model compositional structure in other cognitive domains. Language (Chomsky, 1965) and object perception (Biederman, 1987) have long traditions of emphasizing compositionality. More recently, these ideas have been extended to other domains such as concept (Feldman, 2000) and rule learning (Goodman, Tenenbaum, Feldman, & Griffiths, 2008). Our results add to these attempts by linking compositional function representation to linguistic communication.
There are four important limitations of the current work, which point the way toward future research. First, we do not have a computational account of how patterns are encoded into natural language. Based on work in machine learning (Lloyd et al., 2014), one starting point is to assume that people first infer a structural description of the pattern, and then “translate” this structural description into natural language. Although the work of Lloyd et al. (2014) shows how to do this for the compositional GP model, the natural language descriptions are highly technical, and therefore a rather poor match for lay descriptions of patterns. As the word frequencies in Figure 3a–b illustrate, people seem to make use of more metaphorical language when describing compositional functions—a property not captured by the austere statistical descriptions of Lloyd and colleagues. What we need is a kind of pattern “vernacular” that maps coherently (though perhaps approximately) to the structural description.
The second limitation of our work is that we do not have a computational account of how descriptions are decoded into patterns for redrawing. One natural hypothesis is that this is essentially a reverse of the process described above: natural language descriptions are first translated into structural descriptions, which can then be plugged into the GP model to generate the mean function or sample from the posterior.
Both of these limitations might be addressed in a data-driven way by using machine learning tools to find invertible mappings from structural descriptions to natural language. In particular, we could treat this as a form of structured output prediction, a supervised learning problem in which the inputs and outputs are both multidimensional. Modern structured output prediction algorithms have developed a variety of ways to exploit the structured nature of linguistic data (e.g., Daumé, Langford, & Marcu, 2009; Tsochantaridis, Joachims, Hofmann, & Altun, 2005). These algorithms have not yet been applied to human pattern description.
The third limitation of our work is that we have investigated a fairly small set of functions. This set was chosen based on our past work (Schulz, Tenenbaum, et al., 2017) so as to minimize low-level perceptual confounds. However, further work will be required to verify that our results generalize to a broader range of functions.
The final limitation is that it is currently hard to draw a clear distinction between compositional and noncompositional patterns. Given that both the compositional and the non-compositional model can capture almost any pattern given enough data, the main differences between the two models can be derived from their predictions under a finite data regime. The two models’ inductive biases differ substantially given the number of data points we have applied here. Take as an example patterns that exhibited a linear trend. Even though the noncompositional kernel could eventually capture a linear trend, it would require a large number of noncompositional parts to interpolate trends and yet would still struggle to extrapolate beyond the encountered data; this is because it lacks the required inductive biases to express trends efficiently.
The idea that concepts are represented in a “language of thought” is pervasive in cognitive science (Fodor, 1975; Piantadosi, Tenenbaum, & Goodman, 2016), and we have previously shown that human function learning also appears to be governed by a structured “language” of functions (Gershman, Malmaud, & Tenenbaum, 2017; Schulz, Tenenbaum, et al., 2017). Specifically, people decompose complex patterns into compositions of simpler ones, ultimately producing a structural description of patterns that allows them to effectively perform a variety of tasks, such as extrapolation, interpolation, compression, and decision making. The results in this article suggest that the availability of a structural description can also be used to communicate patterns in natural language. Because noncompositional functions are less effectively encoded into a structural description, they are disadvantaged in terms of accurate pattern communication. This finding provides new insight into how a language of thought might mediate translation between vision, language, and action.
ES received funding from the Harvard Data Science Initiative.
ES: Conceptualization: Equal; Formal analysis: Lead; Investigation: Equal; Visualization: Lead; Writing - Original Draft: Equal. FQ: Conceptualization: Equal; Data curation: Supporting; Software: Lead; Visualization: Supporting; Writing - Original Draft: Supporting. SJG: Conceptualization: Equal; Data curation: Supporting; Supervision: Lead; Writing - Original Draft: Equal.
The authors thank Matthias Hofer for helpful discussions.
Note that although the spectral kernel could a priori be captured by sums of radial basis function (RBF) and periodic kernels, the extracted (i.e., fitted) “human kernel” reported by Wilson et al. (2015) was more similar to a mixture of a radial basis function and a linear kernel. We compare both of these types of mixture kernels to our full compositional kernel in our lesioned model comparison in the Supplemental Materials.
All descriptions can be found online: https://ericschulz.github.io/comcompresps.pdf.
We did not include a random slope over participants into this model comparison, because there was no evidence for a random slope improving model fits for the regression focusing on RBF-specific words, BF = 0.02, the regression focusing on linear-specific words, BF = 0.08, as well as the regression focusing on periodic-specific words, BF = 0.02.
Competing Interests: The authors declare they have no conflict of interest.