Abstract
Decisionmaking behavior is often characterized by substantial variability, but its source remains unclear. We developed a visual accumulation of evidence task designed to quantify sources of noise and to be performed during voluntary head restraint, enabling cellular resolution imaging in future studies. Rats accumulated discrete numbers of flashes presented to the left and right visual hemifields and indicated the side that had the greater number of flashes. Using a signaldetection theorybased model, we found that the standard deviation in their internal estimate of flash number scaled linearly with the number of flashes. This indicates a major source of noise that, surprisingly, is not consistent with the widely used 'driftdiffusion modeling' (DDM) approach but is instead closely related to proposed models of numerical cognition and counting. We speculate that this form of noise could be important in accumulation of evidence tasks generally.
https://doi.org/10.7554/eLife.11308.001eLife digest
Perceptual decisionmaking, i.e. making choices based on observed evidence, is rarely perfect. Humans and other animals tend to respond correctly on some trials and incorrectly on others. For over a century, this variability has been used to study the basis of decisionmaking. Most behavioral models assume that random fluctuations or 'noise' in the decisionmaking process is the primary source of variability and errors. However, the nature of this noise is unclear and the subject of intense scrutiny.
To investigate the sources of the behavioral variability during decisionmaking, Scott, Constantinople et al. trained rats to perform a visual 'accumulation of evidence' task. The animals counted flashes of light that appeared on either their left or their right. Up to 15 flashes occurred on each side, in a random order, and the rats then received a reward if they selected the side that the greatest number of flashes had occurred on. The rats chose correctly on many occasions but not on every single one.
Using a computercontrolled rat training facility or 'rat academy', Scott, Constantinople et al. collected hundreds of thousands of behavioral trials from over a dozen rats. This large dataset provided the statistical power necessary to test the assumptions of leading models of behavioral variability during decisionmaking, and revealed that noise grew more rapidly with the number of flashes than previously predicted. This finding explained patterns of behavior that previous models struggled with, most notably the fact that individuals make errors even on the easiest trials. The analysis also revealed that animals maintain two separate running totals – one of stimuli on the left and another of stimuli on the right – rather than a single tally of the difference between the two.
Scott, Constantinople et al. further demonstrated that rats could be trained to perform this task using a new system that enables functional brain imaging. The next step is to repeat these experiments while simultaneously recording brain activity to study the neural circuits that underlie decisionmaking and its variability.
https://doi.org/10.7554/eLife.11308.002Introduction
Subjects performing perceptual decisionmaking tasks display a large amount of trialtotrial behavioral variability. Determining the sources of this variability could provide insight into the neural mechanisms of decisionmaking and produce more accurate predictions of behavior. It has been proposed that behavioral variability is caused in part by noise accruing during the process of evidence accumulation. This noise may have a variety of origins depending on the behavioral task. It can be inherent in the natural world, produced by the signal detection limits of sensory organs themselves (Barlow and Levick, 1969), or it may reflect the variability of neural responses in the brain at different stages of processing. Previous studies have attempted to trace the sources of noise using a combination of behavioral and neurophysiological approaches.
At the behavioral level, many models of the evidence accumulation process are known as ‘driftdiffusion’ models (DDMs), because a major component of the noise is modeled as diffusion noise in the accumulator, i.e., noise that is independent for each time point (Bogacz et al., 2006; Smith and Ratcliff, 2004, but see Zariwala et al., 2013). A recent study by Brunton and colleagues (Brunton et al., 2013) developed an accumulation of evidence task, which they modeled with a DDMlike process, to isolate different sources of noise. In their task, evidence was delivered in randomly timed but precisely known pulses: two randomly generated streams of discrete auditory clicks were presented from left and right speakers, and subjects (rats and humans) were trained to indicate the side with the greater number of clicks. The precise timing of the stimuli combined with large numbers of behavioral trials enabled fitting of a detailed and statistically powerful behavioral model. The results suggested that in most subjects the diffusion noise in the accumulator was essentially zero (noiseless), and the behavioral variability was best explained by noise that was added with each pulse of evidence. However, it was assumed that each pulse introduced independent noise. This implied that although timelocked to the stimulus, the noise associated with the stimulus was diffusionlike, and the standard deviation of the total stimulusinduced noise scaled as the square root of the number of pulses. Overall, the model was qualitatively consistent with the widelyassumed diffusionlike noise in the evidence accumulation process. However, similar to other reports, diffusionlike noise alone was unable to fully predict behavioral variability, and specifically failed to account for errors on the easiest trials. Following common practice, Brunton et al. (2013) included an additional noise parameter, termed lapse rate, which described variability that did not depend on the stimulus or trial duration. The need for the lapse rate term suggests the existence of additional sources of noise that are not diffusionlike.
To better characterize sources of noise during perceptual decisionmaking, here we developed a visual analogue of the Brunton et al. accumulation of evidence task with two key features (Erlich et al., 2013, SFN, abstract). First, rats could perform the task during voluntary headrestraint (Girman, 1980; 1985; Kampff et al., 2010, SFN, abstract), allowing for the potential of cellular resolution imaging (Scott et al., 2013) and perturbation in future studies (Rickgauer et al., 2014). Second, the sensory stimuli were designed to attempt to isolate internally generated noise by minimizing noise inherited from the stimulus. In our task, rats were presented with two series of brief (10 ms), pseudorandomly timed LED flashes to the left and right visual hemifields. Following the form of the Brunton et al. (2013) task, the side with the greater total number of flashes indicated the location of a water reward. However, whereas the acoustics of the behavior chamber could distort localization of auditory stimuli, here our visual stimuli were presented from wellseparated LEDs positioned to the left and right visual hemifields, suggesting that the rats would have no difficulty distinguishing the left versus right origin of each flash. Additionally, low numbers of flashes were presented in wellseparated time bins.
Behavioral analyses indicated that rats solved the task by accumulating all available evidence, and that flashassociated noise was the predominant source of behavioral variability. We implemented a signal detection theorybased model to evaluate the assumption that noise in the accumulator value was diffusionlike, i.e. that noise was independent for each flash and the noise (standard deviation) of the accumulator value scaled as the square root of the number of flashes. The signal detection theorybased model was designed to find the scaling relationship between noise in the accumulator value and number of flashes that best described the data. Surprisingly, model fits revealed that the noise in the subjects’ numerical estimates scaled linearly with the number of flashes presented, not as the square root. This relationship, called scalar variability, has previously been observed in tasks that require subjects to estimate the duration of a stimulus (Gibbon, 1977) or count the number of stimuli (Gallistel and Gelman, 2000). Moreover, a behavioral model implementing scalar variability predicted imperfect behavioral performance on the easiest trials, without the need for a nonzero ‘lapse rate’ parameter. Linear scaling of the standard deviation of the noise with the number of flashes reveals a source of noise that does not treat individual pulses independently. Therefore, the noise is not diffusionlike; unlike a lapse rate it depends on total number of evidence pulses; and it is introduced after singleflash sensory processing. We suggest that taking into account this form of noise will be a critical factor for understanding variability in decisionmaking behaviors.
An orthogonal approach for measuring noise in the accumulation process is to relate the variability of neuronal responses to behavior (Shadlen et al., 1996; Cohen and Newsome, 2009; Mazurek et al., 2003). Ideally, this approach would include recordings from multiple neurons across brain regions and precise neural circuit manipulations. To this end, we developed a version of our task that rats could perform during voluntary head restraint. Voluntary head restraint was developed as an alternative to forced head restraint in which the initiation and termination of the restraint period are under the control of the animal (Girman, 1980; 1985; A.R. Kampff et al., 2010, SFN, abstract). Recently we reported that rats can be trained to perform voluntary head restraint in a highthroughput semiautomated behavioral facility and described a voluntary head restraint system that provides the stability needed for in vivo cellular resolution imaging (Scott et al., 2013). Here we show that performance of head restrained rats was essentially identical to the performance of rats trained on the unrestrained version of the task. These results demonstrate that rats can perform complex cognitive behaviors during voluntary head restraint and provide a platform for characterizing noise during decisionmaking across multiple brain regions.
Results
In our rat visual accumulation of evidence task (Figure 1A and B), subjects initiate a trial by inserting their nose into the center port of a threeport operant conditioning chamber. Subjects must keep their nose in the center port (fixation) for 1–8 s while a series of brief (10 ms) flashes are presented by LEDs to the left and right visual hemifields. Between 0 and 15 flashes are presented independently to each hemifield with a randomized number and timing of flashes on each side (Figure 1—figure supplement 1). Trial durations and number of presented flashes are independently varied on a trialbytrial basis. Following an auditory go cue that signals the end of fixation, subjects receive water reward (typically 25 μL) for orienting to a nose port on the side that had more flashes.
We trained seven rats to perform this task using an automated procedure in a highthroughput behavioral facility. Rats progressed through a series of stages in which they learned (1) to associate light with reward, (2) to maintain nose in center fixation for increasingly long durations and (3) to compare the number of flashes on each side to predict the rewarded location. The procedural code used for training can be downloaded from http://brodylab.org/code/flashcode. Rats progressed through the training stages in 3500 trials. Fully trained rats performed a combined 525,073 trials at 91% for the easiest trials and 70% correct overall. Moreover rats were sensitive to the difference in the number of flashes including differences of a single flash (Figure 1C and D; Figure 1—figure supplement 2).
Errors in behavioral choice increase with number of flashes, not trial duration
Accumulation of evidence involves two processes: maintaining a memory of the evidence and adding new evidence to that memory. To assess whether noise (and thus behavioral variability) was more closely associated with the memory of the accumulator or with incoming sensory evidence, we initially fit the Brunton et al. model to our visual task data. This produced results consistent with those found for Brunton et al.’s auditory task, including nearzero estimates of accumulator memory noise (i.e., a predominant role for incoming sensory evidence noise), and long accumulation time constants (Figure 2—figure supplement 1). However, further analysis described below led us to question Brunton et al.’s assumption of independent noise across pulses of sensory evidence. We therefore took a modelfree approach to estimate whether noise was more closely associated with the memory of the accumulator or with incoming sensory evidence. When trial duration and flash difference were held constant, errors increased with total number of flashes presented, suggesting that noise increased with each flash (Figure 2A). Next, we sought to directly compare the effects of flashes and time on behavioral performance (% correct). First, looking across trials with identical differences in flash number (fixed #R#L = ΔF) but with varying total flash number (#R+#L=ΣF) , we calculated the fraction of correct responses as a function of ΣF, relative to the average performance (Δ Performance; Figure 2C). With the difference in flashes ΔF thus controlled for, we found that trials with greater numbers of flashes showed a substantial decrease in performance. Linear regression suggested that each additional flash decreased performance by 1.14% (+/0.1%). With an average of 4 flashes presented per second, increasing total flashes at a fixed flash difference thus produced an average decrement in performance of 4.56% per second. Then, to estimate the effect of time on performance, we calculated Δ performance across trials with both identical flash differences (ΔF) and identical total flash number ΣF), but with different overall trial duration. This analysis, which controls for the number of flashes, revealed that the purely timedependent decrement in performance was only 0.037% (+/. 77%) for each second of trial duration. These results suggest that error rates depend far more strongly on the number of flashes than on trial duration. We found this both for the group data of Figure 2 and for individual rats as well (Figure 2—figure supplement 2; Figure 2—figure supplement 3).
To assess whether the rats exhibited long accumulation time constants we next computed the ‘psychophysical reverse correlation’ (Brunton et al., 2013; Kiani et al., 2008; Nienborg and Cumming, 2009 ; see Materials and methods). For each time bin, we computed the probability that there was an excess flash in that bin on the side to which the subjects subsequently oriented. This analysis indicated that flashes across all time bins contribute equally to subjects’ decisions, both when data were pooled across rats (Figure 2B), and for individual rats (Figure 2—figure supplement 4). This result suggests accumulation time constants that are longer than the trial duration, since subjects’ choices were equally influenced by early, middle, and late flashes.
Taken together, behavioral analyses and behavioral model fits suggested that rats based their choices on evidence accumulated over the entire trial duration and that the noise associated with each flash was the predominant source of noise in the accumulation process.
Subjects’ estimates of flash number exhibit scalar variability
As described in the introduction, in the drift diffusion framework assumed in previous behavioral models of evidence accumulation, the standard deviation of the total flashassociated noise scales as the square root of the total number of flashes (Brunton et al., 2013). However, two popular alternative models, called scalar variability and subitizing, have been proposed in which noise scales differently with evidence. In the scalar variability model, which has been used to describe both time estimation and counting, the standard deviation of the estimate shows a linear relationship with the quantity represented. The subitizing model predicts that the numerical representation of the first few numbers, up to three or four, is essentially noiseless, and after five its standard deviation grows linearly with the number of flashes. Subitizing has been proposed in tasks in which stimuli are presented simultaneously (Trick and Pylyshyn, 1994) and also sequentially (Camos and Tillmann, 2008; von Glasersfeld, 1982).
To quantify the amount of noise associated with different numbers of flashes, we used a signal detection theorybased framework that abstracted away the sequential nature of the stimulus presentation, and focused instead on the total number of flashes on each side, and noise associated with those totals, as the key determinants of the animal’s decisions (Figure 3A). In this approach, the subjects’ estimate of the number of flashes on one side is modeled as a random variable drawn from a Gaussian distribution. For n flashes presented on a side, the mean of this distribution is equal to n, and the variance is a free parameter, ${\sigma}_{n}^{2}$. If there can be up to 15 flashes presented on a side, then, there will be 16 free parameters in the model, ${\sigma}_{0}^{2}$ through ${\sigma}_{15}^{2}$. On each trial the subject selects two such random variables, one for the total number of flashes on the left and one for total flashes on the right, compares the two variables and orients to the side for which the random sample was greater. Correct responses occur when the random sample associated with the side that had the greater number of flashes is greater than the random sample from the other side.
We used maximum likelihood estimation to calculate the standard deviations of each distribution (${\sigma}_{0}...{\sigma}_{15}$) given the number of flashes and behavioral choices across all trials. We found that the bestfit standard deviation values scaled approximately linearly with the number of flashes (Figure 3B; Figure 3—figure supplement 1). We next compared the predictions of the model with the behavioral data (Figure 3C and D). The model was able to capture a large number of features of the data, including the subjects’ imperfect performance on the easiest trials, even though no lapse rate parameter was included in the model (Figure 3D). This was true both for the group data in Figure 3 and for individual rats (Figure 3—figure supplement 2). Note, although our model represents numerical estimates as a scalar random variable on each trial, our results are also consistent with probabilistic numerical representations in the brain (Kanitscheider et al., 2015).
We used least squares regression combined with bootstrapping to estimate the bestfit scaling of ${\sigma}_{n}$ as a function of n according to each of the three models of noise: scalar variability (SV, linear scaling), subitizing (SUB +SV, constant at zero up to some value of n, then linear scaling) and linear variance (LV, square root scaling) (Figure 3E). Comparison of goodnessoffit value (r^{2}) using a nonparametric permutation test confirmed that the MLE standard deviation estimates (${\sigma}_{0}...{\sigma}_{15}$) were better fit by the scalar variability model than the subitizing or linear variance models (Figure 3F; see Materials and methods: Model Comparison).
Scalar variability suggests that scaling the number of flashes on each side by the same factor should lead to identical discriminability: if ${\sigma}_{n}=k\times n$, then the probability of choosing right (equation in Figure 3A) is a function of the ratio r = N_{R}/N_{L}, where N_{R} and N_{L} are the number of flashes on right and left, respectively. For the easiest trials (e.g., N_{L} >> 1 and N_{R} = 0), performance asymptotes at
producing a nonzero lapse rate even without an explicit lapse rate parameter (see mathematical appendix, Appendix 1). To test the prediction that performance should be constant for trials with a fixed ratio r, we compared behavioral performance on trials with constant ratios of flashes on the rewarded and unrewarded sides. Supporting scalar variability, both the data and bestfit signal detection theorybased model predictions exhibited roughly constant performance for trials with fixed ratios of flashes (Figure 3—figure supplement 3). Interestingly, however, behavior deviated from scalar variability, especially for trials with very low number of flashes. For these trials, there was an improvement in performance with additional flashes, even though the ratio of flashes was fixed (Figure 3—figure supplement 3), which is not predicted by pure scalar variability. Nevertheless, this deviation from scalar variability was well predicted by the model parameters of Figure 3B. It is thus likely driven by the deviation from pure linear dependence of the ${\sigma}_{n}$ parameters on n for n<2 (Figure 3B).
We repeated the signal detection theorybased analysis on behavioral data collected from rats performing a previously developed auditory accumulation of evidence task (Brunton et al., 2013; Hanks et al., 2015). We accounted for adaptation effects, which are prominent in the auditory task due to high click rates, and computed the ‘effective’ number of clicks presented after implementing the adaptation dynamics described in Equation (6) (see Materials and methods). On trials in which this approach yielded fractional number of clicks, we rounded the effective click number (after adaptation) to the nearest integer. This kept the number of parameters in the auditory task model roughly equal to the number in the model for the visual task. We then used maximum likelihood estimation to calculate the standard deviations of each distribution (${\sigma}_{0},\dots ,{\sigma}_{n}$) given the effective number of clicks and the subjects’ behavioral choices. Consistent with our results with the visual flashes task, the goodnessoffit value (r^{2}) revealed that the MLE standard deviation estimates in the auditory task were better fit by a model of scalar variability plus a constant offset (SV, σ_{n}= ${k}_{0}+k\times n$) compared to the linear variance previously assumed by sequential sampling models (LV, σ_{n}^{2} =${k}_{0}+k\times n$) (r^{2} = 0.97 for SV vs r^{2} = 0.89 for LV, p << 0.001 based on nonparametric permutation test, Figure 3—figure supplement 4, 5). Our findings are thus not specific to the visual modality. Notably, perfect scalar variability (${\sigma}_{n}=k\times n$ without the constant ${k}_{0}$) would predict no improvement as a function of time for a fixed click ratio, yet such an improvement is clearly seen in the clicks data (Figure 3—figure supplement 6). The SV model accounted for this by having a significantly nonzero offset, ${k}_{0}$, and was thus able to fit the clicks data quite well (Figure 3—figure supplement 7). A nonzero fixed offset, ${k}_{0}$, thus represents a source of noise that is added to the accumulators, independently of the stimulus that was presented. Notice that this source of noise does not account for the lapse rate, since ${k}_{0}$ does not appear in the expression for the lapse rate (see mathematical appendix, Appendix 1). A larger offset (${k}_{0}$) was the main qualitative difference we found between fitting the auditory clicks data (for which ${k}_{0}$ = 2.10 +/ 0.20, mean +/ s.e.m. across rats) and the visual flashes data (for which ${k}_{0}$ = 0.49 +/ 0.142).
Comparison of behavioral models suggests the presence of at least two accumulators
The signal detection theorybased model described above makes two assumptions about the nature of the accumulation process. First it assumes that subjects maintain separate estimates of left and right flashes during a behavioral trial. Second it assumes that noise depends only on the final value of the left and right integrator. However, many quantitative models of accumulation of evidence propose that subjects maintain a single decision variable (e.g., a running estimate of the difference in number of right minus left pulses; Bogacz et al., 2006). Other binary decision models have proposed two separate accumulators in which noise is added throughout the accumulation process (Ratcliff et al., 2007; Usher and McClelland, 2001).
Therefore, we decided to evaluate the assumption of a dual accumulator and the assumption that the noise depends only on the final value of the integrator(s), by comparing the performance of multiple variants of our signal detection theorybased behavioral model (Figure 4). We probed three independent questions, the combined answers to which produced eight different models: (1) Is performance best fit by dual accumulators or by a single accumulator? (2) Is performance best fit by gradual accrual of noise throughout stimulus presentation, or by noise that depends only on the final value of the integrator? And finally, as in the previous section, (3) Is variability best fit as standard deviation being linear in flash number, or by variance being linear in flash number? Each model (ah) had two free parameters β_{1 }and β_{2 }that determined the relationship between the number of flashes, n, and the standard deviation of the flash number, ${\sigma}_{n}$ (Figure 4B). Four models (a,b,e,f) assumed scalar variability (σ = β_{1}*n+β_{2}) and four (c,d,g,h) assumed linear variance (σ^{2 }= β_{1}*n+β_{2}). Four models assumed a single accumulator (a,c,e,g) and four assumed independent left and right accumulators (b,d,f,h). Four models (ad) assumed that noise was based only on the final estimate of the accumulator (static sampling), and four assumed (eh) that noise was added to the estimate of the accumulator at the time of each flash (sequential sampling).
We used maximum likelihood estimation to find, for each model, the values of β_{1} and β_{2} that best fit the behavioral data. Despite the closely related structure of the eight models, and having precisely the same number of free parameters, the likelihood of the data at the bestfit parameter values for nearly all of the twoaccumulator versions (b,d,h) of the model was significantly greater than the corresponding oneaccumulator versions (a,c,g) (p<.001; nonparametric test using bootstrapping; see Materials and methods), indicating better performance for the twoaccumulator model (Figure 4D,E). Overall we found that the static, dual accumulator scalar variability model had the highest likelihood across all models (p<.001; nonparametric test using bootstrapping see Materials and methods: Model Comparison; Figure 4—figure supplements 1, 2). This suggests that the main noise source, in the limited regime of models considered here, operates on the final value of the accumulators, rather than being incrementally and sequentially added to the accumulators as the process unfolds. We emphasize, however, that we have not made an effort to systematically explore the full space of possible models in which noise is added gradually during the stimulus presentation, so a conclusion that noise is added only to the final value of the accumulator remains tentative.
Subjects accumulate flash number, not flash duration
Scalar variability has been proposed both for numerical and interval timing estimation, raising the possibility that rats could be accumulating the summed duration of the flash ‘ON’ times. Alternatively, it is possible that rats could be treating each flash as a distinct event, independent of its duration, and integrating the number of flashes. To test which strategy the rats used, we performed an additional experiment in which we trained a second cohort of rats (n=7) on a new version of the task. As before, rats were rewarded for orienting to the side with the greater number of flashes, but in this version the durations of each flash and interflash interval were randomly jittered (see Materials and methods). Flash duration was drawn from a Gaussian distribution with a mean of 10 ms and standard deviation of 1.5 ms (Figure 5A,B). Because LED duration is correlated with number of flashes on most trials, it is difficult to evaluate whether animals use flash duration or flash number on most trials (Figure 5C,D). However, on trials in which there was no difference in the number of flashes (Δflashes = 0), the jittering of flash duration led to the generation of some trials with a difference in the duration presented from the right or left LEDs (Figure 5B). Since, for example, 20 ms of light is the equivalent of a difference of two flashes, we reasoned that on these trials, if the subjects were integrating light duration, they would be more likely to orient to the side with longer LED durations, and a regression line fit to the data in Figure 5E should have a positive slope. Conversely, if they were integrating discrete flash number, they should not exhibit a preference to the side with longer flash duration, and a regression line fit to the data in Figure 5E should have a slope of zero. Consistent with integrating flash number, we found that on trials with 0 Δflashes, rats did not exhibit a preference for the side with longer flash durations (Figure 5E,F). These results indicate that the rats were accumulating the number of flashes, not their total duration.
Reward and error history bias future behavior choices
One potential additional source of variability in the rats’ behavior is the location of reward on previous trials. In some behavioral tasks, subjects display a winstayloseswitch approach to decisionmaking, in which choices that lead to reward tend to be repeated, while alternative unrewarded choices tend to be abandoned. To assess whether the memory of reward location on previous trials biased the subjects’ decisions on future trials, we identified trials following a reward and compared behavioral performance depending on whether the reward had been delivered to the left or right (Figure 6A). On these trials subjects continued to perform the accumulation task, but were biased toward the side where they had previously obtained reward. The effect of this bias was modest but significant, and was constant across a range of numerical flash differences, producing a vertical shift in the psychometric function (Figure 6A). Similarly, if the subjects made an error on one side, they exhibited a bias toward the other side on the subsequent trial (Figure 6B).
Next, we extended this analysis to determine how long this bias persists. We computed the probability that subjects chose to return to the same side where they had previously obtained a reward and repeated this computation for progressively later choices (Figure 6C, black line). The behavioral side bias caused by a reward decreased steadily over the next three trials and after three trials no significant bias was observed. We repeated this analysis for error trials and found that the bias was smaller in magnitude, but displayed the same steady decrease for three trials into the future (Figure 6C, gray line). After three trials no significant bias was observed.
Finally, when rewards were consecutive (i.e. more than two correct responses to the right in a row) the bias observed was additive. The side bias observed from two consecutive rewards on the same side is not significantly different from the sum of the reward biases from one and two trials back (Figure 6D). Together these results suggest that (1) both rewards and errors bias the subject’s choices up to three trials in the future and (2) reward history effects combine linearly to influence the behavior of the subject. Similar results have been observed in the auditory, Brunton et al., version of the accumulation of evidence task (Chan, Brunton, and Brody, unpublished data).
Performance of voluntarily head restrained rats on a visual accumulation of evidence task
Future studies into the biological mechanisms of decisionmaking will be facilitated by the ability to perform precise neuronal circuit perturbations and cellular resolution imaging. Behavioral tasks that can be performed during head restraint allow the experimenter to perform techniques such as cellular resolution optical stimulation and largescale twophoton calcium imaging that are difficult to employ in unrestrained preparations. Therefore we developed a version of the visual accumulation of evidence task that could be performed during voluntary headrestraint.
In the headrestrained version of the task, subjects initiate a behavioral trial by guiding a surgically implanted titanium headplate into a custom headport mounted on one wall of an operant training chamber (Figure 7A, left and 7B). The trial begins when the leading edge of the headplate contacts two miniature snap action switches (contact sensors) mounted on the headport. Voltagecontrolled pneumatic pistons then deploy to immobilize the headplate and, using the principles of kinematic mounts, register the head to within a few microns. During the restraint period, which lasts 2–3 s, headrestrained animals are presented with up to six flashes on each side (Figure 7A, middle). After a brief (500 ms) delay, subjects are released from restraint, and can obtain water reward by orienting to one of two side ports mounted to the wall flanking the headport (Figure 7A, right). Headrestrained subjects can also choose to terminate the restraint period early by operating a release switch located on the floor of the chamber. Trials that are terminated early result in a brief timeout (~2 s) in which no reward can be obtained and no trial can be initiated.
To compare behavioral performance to unrestrained rats, we trained seven rats on the headrestrained version of this task. The initial stages of training were performed in a highthroughput facility, while acclimation to the pistons and data collection from fully trained animals were performed in a separate facility with a single dedicated behavioral chamber. Rats completed the initial stages of training in similar time to the unrestrained rats. However, a portion of rats (~40%) were slower to acclimate to the pistons than the remainder of their cohort and were excluded from further study.
Behavioral analysis revealed comparable performance between voluntarily headrestrained and unrestrained rats (Figure 7C and D). Fitting the accumulator model produced model parameters in the headrestrained subjects that was similar to the best fit parameters for the unrestrained rats: negligible accumulator noise, predominant flashassociated noise, accumulation time constants longer than the trial duration and bounds higher than the number of possible flashes. Fitting the signal detection theorybased model revealed similar scaling of noise between restrained and unrestrained versions of the task (Figure 7E). We computed the ‘psychophysical reverse correlation’ to further assess whether the rats exhibited long accumulation time constants (see Materials and methods). For each time bin, we computed the probability that there was an excess flash in that bin on the side to which the subjects subsequently oriented. This analysis indicated that flashes in each time bin contribute equally to subjects’ decisions (Figure 7F). This was true for both voluntarily headrestrained and unrestrained rats.
Discussion
In this study we developed a visual accumulation of evidence task to study the sources of noise that contribute to behavioral variability. Subjects were presented with two simultaneous trains of randomly timed flashes, one train on their left, and the other on their right, and were rewarded for orienting towards the side with the greater total number of flashes. Behavioral analyses, including reverse correlation, indicated that rats used flashes presented throughout the entire trial to guide their decision, consistent with a strategy of accumulating evidence (Figure 2—figure supplement 4). We used a signal detection theory framework to estimate the noise associated with each pulse of evidence (flash). This analysis showed that noise (standard deviation) in the subjects’ numerical estimates scaled linearly with the number of flashes presented, indicating that noise per flash is not independent across flashes. We find that noise is also independent of total stimulus duration, and is perhaps added to the decision variable after the accumulation process. Our behavioral modeling further suggests the existence of two separate accumulators in this binary decision task (Figure 4). The noise scaling we found is similar to scalar variability described in the numerical cognition literature, and appears to be the major source of variability in our subjects’ behavior. Yet it is not diffusionlike and is therefore not welldescribed by the majority of current ‘driftdiffusion modeling’ (DDM) approaches for accumulation of evidence in decisionmaking. Incorporating this scalar variability noise will likely be critical for understanding variability in decisionmaking behaviors. Finally, we demonstrate that rats can be trained to perform this task during voluntary head restraint, which opens up new experimental avenues for uncovering the circuit mechanisms that generate behavioral variability during decisionmaking.
Performance of cognitive behaviors during voluntary head restraint
Headrestraint is an experimental tool widely used in psychology and neuroscience to immobilize the head of a subject. The technique is typically employed in behavioral experiments to facilitate reliable presentation of sensory stimuli to the head or face and precisely control the subjects’ movements, and in neurophysiological experiments to minimize motion between the brain and recording apparatus. However, it can take some species, notably rats and primates, weeks to acclimate to forced head restraint. Moreover forced head restraint requires skilled experimenter intervention and is difficult to incorporate into automated training procedures. Voluntary head restraint has been proposed to reduce the stress normally associated with forced head restraint and facilitate training on sophisticated behavioral tasks (Scott et al., 2013). Here, we show that rats are able to perform a complex cognitive task, involving both numerical representation and integration of evidence, while voluntarily headrestrained. This provides a foundation for identifying and characterizing the neural circuit mechanisms of accumulation of evidence and numerical representation in future studies using in vivo cellular resolution imaging and perturbation (Rickgauer et al., 2014).
Sources of noise in accumulation of evidence tasks
The behavioral task described here exhibited a number of features that enabled detailed characterization of noise during accumulation of evidence. The visual stimuli presented in this study were: (1) low in number; (2) presented reliably to left and right visual hemifields; (3) presented from wellseparated light sources in front of the rat’s face; (4) presented at sufficient intervals to avoid adaptation or facilitation between successive stimuli. All of these features attempted to minimize sources of noise other than noise in the subjects’ internal estimates. Consistent with previous reports (Brunton et al., 2013; Kiani et al., 2013), we found that behavioral variability was not affected by the trial duration but was strongly correlated with the number of flashes presented.
However, we identified a number of additional sources of noise. First, rats exhibit a modest side bias due to reward history occurring up to three trials in the past (Figure 6). In dynamic foraging tasks, strategies that incorporate trial history (such as winstay, loseswitch strategies) often maximize rewards (Corrado et al., 2005; Herrnstein, 1961; Sugrue et al., 2004). However, in the visual accumulation of evidence task, the rewarded side is assigned independently on each trial, so trial history adds noise to the decision process. The influence of trial history on subsequent decisions has been observed in a number of other studies, suggesting that it is a general source of behavioral variability that should be incorporated into accumulation of evidence models (Busse et al., 2011; Gold et al., 2008).
We also found that the standard deviation in the accumulated estimate scaled linearly with the number of flashes. This type of noise scaling, called scalar variability, was consistent across the range of stimuli presented (115 flashes, with minor deviations at the lowest number of flashes) suggesting that the subjects did not use a subitizing approach, i.e., they did not use rapid accurate counting for numbers less than five. We applied our analysis to a closely related previous auditory task that, as in our current visual task, also delivered sensory evidence in discrete pulses (Brunton et al., 2013). We found that our conclusions also applied to the auditory data, with the standard deviation of the noise after N flashes welldescribed by σ_{n}= ${k}_{0}+k\times n$. Our observation that performance in the auditory task improves with increased integration times is qualitatively consistent with data from tasks in which evidence is delivered in a continuous stream rather than in discrete pulses. This leads us to suggest that scalar variability should be an important consideration in the analysis of data from continuous tasks. But whether or not scalar variability is the dominant form of noise in continuous stream tasks, as it was in the pulsatile tasks analyzed here, remains to be determined.
Implications for neural mechanisms of perceptual decisionmaking
The primary focus of this work has been to provide a quantitative description of behavior during accumulation of evidence and to evaluate the sources of behavioral variability. However, an important distinction should be made between quantitative models of behavior and mechanistic models of the underlying neural processes (Marr, 1982). Our results do not suggest an underlying neural mechanism, however they do place important constraints on such mechanistic models. One such constraint is scalar variability, which has been observed in a wide variety of tasks requiring animals to estimate magnitudes (such as time or number Gallistel and Gelman, 2000; Gibbon, 1977). Two possible neural mechanisms have been proposed to explain this phenomenon. The first is that the brain represents magnitude on a logarithmic scale with constant noise (Fechner, 1860; Dehaene and Changeux, 1993; Meck and Church, 1983; Nieder and Dehaene, 2009). Neurophysiological recordings, particularly in sensory systems, have provided support for this mechanism. For example, neurons in the rodent somatosensory cortex exhibit firing rates that scale logarithmically with frequency of tactile stimulation (Kleinfeld et al., 2006). In addition, neurons in primate parietal cortex appear to encode numerosity on a nonlinear scale (power law or logarithmic; Nieder and Miller, 2003). The second possibility is that the brain represents magnitude on a linear scale with noise that is proportional to the magnitude. The results described here did not allow us to differentiate between these two models, and indeed, both models make identical psychophysical predictions (Dehaene, 2001). However, how they would be instantiated in the brain is quite different. For the logarithmic case, the signal diminishes with the quantity being represented, but noise added at each time point is independent. For the linear case, the signal is constant (implying near perfect summation of signals), but noise scales linearly with the magnitude being estimated.
Relation to accumulation of evidence models
Sequential sampling models of accumulation of evidence represent the decision process as the movement of one or more latent decision variables towards a criterion value. Such models have been useful to study a wide range of decision processes. One key advantage is that they provide a momentbymoment estimate of the decision variable on each trial. This timevarying estimate can be compared to the observed neuronal dynamics, allowing analysis of the correlation between cellular firing rates and the subject’s internal representation of evidence (Hanks et al., 2015).
Our results suggest three modifications that could improve the accuracy of sequential sampling models of accumulation of evidence. First we found that reward and error trials introduce a bias that persists up to three trials into the future. Second we found that standard deviation in the decision variable increased linearly with the total number of flashes. Note that this observation is inconsistent with the common assumption made in drift diffusion models, that noise is added independent of time and the value of the decision variable. Third, the data was better fit by models that accumulate left and right flashes separately than by models that accumulate a single decision variable. The existence of two accumulators (Ratcliff et al., 2007; Usher and McClelland, 2001) has been previously observed in the vertebrate brain, for example the oculomotor system in the goldfish contains two lateralized accumulators that integrate transient motor commands to maintain a memory of gaze position (Aksay et al., 2007), and is consistent with electrophysiological recordings in primates (Bollimunta and Ditterich, 2012) and rodents (Hanks et al., 2015).
In the future, it will be interesting to develop a reactiontime version of the task, in which subjects are allowed to determine the time during which they observe the stimulus. This would allow testing predictions from recent pulseaccumulation models (Simen et al., 2015).
Noise as a mechanism of generalization during perceptual learning
Variability in behavior and neuronal networks has been proposed to result, in part, from biophysical constraints of the nervous system. In some systems, however, noise is thought to serve an important functional role. For example it has been proposed that random fluctuations in neuronal membrane potential contributes to contrast invariance in the visual system (Anderson et al., 2000). Moreover, during motor learning, noise is actively introduced into highly reliable systems to create the behavioral variability required for trial and error learning (Olveczky et al., 2005). An interesting question is whether noise also facilitates learning in perceptual tasks. One potential function of noise in the estimate of flash number is that it could allow animals, which have been trained on a reduced stimulus set, to generalize, and perform correctly on trials with numbers of flashes they have never encountered. For example, if rats were only trained on trials with 5 and 6 flashes, because of the variability of their estimate of 5 and 6 flashes, on a portion of those trials they will have effectively perceived 7 and 8 flashes, and could learn the correct response to those trials by trial and error. Thus noise may allow the observer to experience a larger range of stimuli than they actually encountered, enabling generalization.
Conclusion
Highthroughput operant conditioning combined with detailed, quantitative analysis of behavioral variability allowed us to characterize the process of evidence accumulation in rats. This analysis revealed that animals’ choices were likely derived from a circuit with multiple accumulators and that decisions were influenced by noise from multiple sources. The major source of noise displayed ‘scalar variability’ in the sense that its standard deviation scaled linearly with the number of pulses of evidence. This observation is inconsistent with the assumption made by most previous models of accumulation for decisionmaking, namely that the primary noise source is diffusionlike, added to the accumulator independent of time or accumulator value. Our observation is instead consistent with previous models of numerical cognition, and suggests that evidence accumulation and numerical cognition may be subserved by similar or related circuits and neural mechanisms. We speculate that noise that obeys scalar variability could be an important component in models of accumulation of decisionmaking evidence generally. Finally, the observation that voluntarily headrestrained rats exhibit similar behavioral performance to unrestrained rats, and can be trained to perform complex decisionmaking tasks, suggests the possibility of using cellular resolution optical imaging and perturbation technologies to characterize the neural substrates of cognition.
Materials and methods
Animal subjects
Request a detailed protocolAnimal use procedures were approved by the Princeton University Institutional Animal Care and Use Committee (IACUC; Protocols #1837 and #1853) and carried out in accordance with National Institutes of Health standards. All subjects were male LongEvans or Sprague Dawley rats weighing between 200 and 500 g (Taconic, NY). Rats were placed on a water schedule in which fluids are provided during behavioral training. If rats consumed less than 3% of their body weight in water, they received ad lib water for 1 hr.
Behavior
Request a detailed protocolRats progressed through several stages of an automated training protocol before performing the task as described in the results. All data described in this study were collected from fully trained rats. Sessions with fewer than 100 completed trials were excluded from analyses. These sessions were rare and usually caused by hardware malfunctions. From the sessions included in this paper, on average, rats completed 385 trials per session (median: 362 trials per day; range: 128–1005 trials per day), and performed at 70% correct (median: 71%; range: 56–88% correct; see Figure 1—figure supplement 1).
Stimulus generation
Request a detailed protocolThe cue period consisted of fixed time bins, each 250 ms in duration. Unrestrained rats experienced up to 15 bins, whereas headfixed rats experienced up to 6 bins. For a given trial, each of the two LEDs had a fixed generative probability of producing a flash in each bin. In the final training stage for the unrestrained rats, the generative probabilities for the highprobability LED and lowprobability LED ranged from 7080% and 2030%, respectively. For restrained rats, the generative probabilities were 70% and 30%. However, the rats were rewarded based on the number of flashes that were actually presented, not on the underlying generative probabilities. In other words, in the rare event that the LED designated ‘highprobability’ happened to produce fewer flashes than the LED designated ‘lowprobability’, the animal was rewarded for orienting to the side that had more flashes. The flashes were 10 ms in duration (except for an experimental group in which the flash durations were jittered; see Figure 5), and they occurred in the first 10 ms of each cue bin. For the unrestrained rats, the interflash intervals ranged from 240 ms (two flashes in subsequent bins) to 3.45 s (two flashes separated by 13 bins; see Figure 1—figure supplement 1). For the headfixed rats, the interflash intervals ranged from 240 ms to 1.24 s. In a subset of experiments, interflash intervals and flash durations were jittered. In these experiments, flash durations were drawn from a Gaussian distribution with a mean of 10 ms, and standard deviation of 1.5 ms. Interflash intervals were drawn from a Gaussian distribution with a mean of 250 ms and standard deviation of either 25 or 50 ms.
In the unrestrained version of the task, there was a 500 ms precue period before the beginning of the cue period. For the majority of trials, the duration of the postcue (or memory) period was randomly drawn from a uniform distribution from 5 to 500 ms. In a small subset of trials, the postcue period was drawn from a broader distribution ranging up to 6.5 s, but those trials represent a minority of the dataset (see Figure 1—figure supplement 1). In the headfixed version of the task, the precue period was 1 s in duration, to allow time for the pistons to actuate and fix the headplate in place. The postcue period was 500 ms.
Behavioral control system
Request a detailed protocolOperant training chambers were controlled by the freely available, opensource software platform Bcontrol (Erlich et al, 2011; Scott et al. 2013). Bcontrol consists of an enhanced finite state machine, instantiated on a computer running a realtime operating system (RTLinux), and capable of state transitions at a rate of 6 kHz, plus a second computer, running custom software written in MATLAB. Each behavioral trial consisted of a sequence of states in which different actuators—for example, opening of a solenoid valve for water reward—could be triggered. Transitions between the states were either governed by elapsed times (e.g., 40 ms for water reward) or by the animal’s actions, which caused changes to the voltage output of a sensor in the chamber.
Surgery
Request a detailed protocolSurgical procedures for headplate implantation have been previously described (Scott et al., 2013). Briefly, we anesthetized animals with isoflurane in oxygen and gave Buprenorphine as an analgesic. Once anesthetized, the scalp and periosteum were retracted, exposing the skull. Dental cement (Metabond) was used to bond the headplate to the skull. After a 2week recovery period, implanted animals began training in voluntary head restraint.
Psychophysical reverse correlation
Request a detailed protocolThe choiceconditioned reverse correlation analysis reveals the degree to which flashes in each time bin contribute to the decision. We selected trials in the final training stage that had the same generative probabilities (γ) for generating flashes (p(flash on highprobability side) = 0.7; p(flash on lowprobability side) = 0.3). We computed the average number of left ($\overline{f}$_{L}(t)) and right ($\overline{f}$_{R}(t)) flashes in each time bin (t) conditioned on whether the animal went right or left, to obtain the average difference in flash number:
We next computed the expected difference in flash number for each trial (0.7–.3 = 0.4 for correct trials, 0.4 for error trials), and averaged across trials conditioned on choice. We subtracted this number (the average expected difference in flash number in each bin) from the observed average difference in flash number:
We multiplied $\overline{F}$_{went_left} by 1 so that the two vectors ($\overline{F})$ were both in units of excess right flashes. $\overline{F}$_{went_right} and $\overline{F}$_{went_left} are the red and blue lines plotted in Figure 2B. Positive/negative values in $\overline{F}$ are time bins in which more right/left clicks occurred than expected by chance, respectively.
8parameter sequential sampling model
Request a detailed protocolTo obtain a momentbymoment description of the decision process, we implemented a behavioral model fit to the trialbytrial data that has been described previously (Brunton et al., 2013; Hanks et al., 2015). On each trial the model converts the incoming stream of discrete left and right flashes into a scalar quantity a(t) that represents the gradually accumulating difference between flashes presented to the two sides. At the end of the trial, the model predicts whether the animal would go right or left if a is positive or negative, respectively. The rat’s behavior is used to fit parameters that govern how a(t) evolves. These parameters quantify sensory noise and noise associated with time, leakiness/instability of the accumulation process, sensory depression/facilitation, side bias, and a lapse rate that corresponds to a fraction of trials on which a random choice is made. The dynamics of a(t) are implemented by the following equation:
where δ_{t,tR,L} are delta functions at the flash times, η are Gaussian variables drawn from N(1, σ_{s}), dW is a whitenoise Wiener process, and C parametrizes adaptation/facilitation of subsequent flashes. Adaptation/facilitation dynamics of C are implemented by the following equation:
In addition, a lapse rate parameter represents the fraction of trials on which the rat responds randomly.
Signal detection theory model
Request a detailed protocolTo describe how noise in the decision process scales with number of flashes, we implemented a model to estimate the width of the distribution of animals’ internal estimates of flash number. This signal detection theorybased model assumes that on a given trial, the animal’s estimate of flash number presented to each side is a random variable drawn from a Gaussian distribution whose mean is the number of flashes on that side, and whose variance is a free parameter in the model. The difference of those Gaussians predicts the subjects’ performance: correct trials occur when the difference of Gaussians is positive (i.e. the random variable drawn from the distribution representing the larger number is in fact larger than the random variable drawn from the distribution representing the smaller number). This was implemented by the following equation:
Where L and S are the means of the distributions for the larger and smaller numbers, respectively. ${\sigma}_{L}$ is the standard deviation of noise when L stimuli has been shown for the ‘larger/correct’ side, and ${\sigma}_{S}$ is the standard deviation of noise when S stimuli has been shown for the ‘smaller/incorrect’ side. In other words, ${\sigma}_{L}^{2}$ and ${\sigma}_{S}^{2}$ are the variances of the distributions of the numerical estimates of L and S. The values of ${\sigma}_{L}^{2}$ and ${\sigma}_{S}^{2}$ that maximized the likelihood of the animals’ behavioral choices were fit using the Matlab function fmincon. The values of ${\sigma}_{N}^{2}$ reported in this paper were derived using a bootstrapping approach. One thousand surrogate data sets were created by selecting behavioral trials at random with replacement from the original data set. The signal detection theorybased model was then fit to each surrogate data set, producing 1000fold estimates for ${\sigma}_{0}^{2}$ … ${\sigma}_{15}^{2}$. The ${\sigma}_{n}^{2}$ values reported in the paper represent the mean of over these values and the confidence intervals were derived from the standard deviation of these values.
Model comparison
Request a detailed protocolTo evaluate the goodnessoffits of the different models, we implemented a nonparametric permutation test. First, for each model, we computed a distribution of bootstrapped r^{2} values by resampling the behavioral trials with replacement, performing the fitting procedure, and computing the r^{2} value, on 1000 iterations. For pairwise comparisons, (for example, between the scalar variability (SV) and linear variance (LV) models in Figure 3F), the null hypothesis was that the r^{2} values for each distribution derive from a common distribution. To test this hypothesis, we combined the bootstrapped r^{2} values from the SV and LV models into a single distribution, and from that combined distribution, created two arbitrary distributions of fake SV and LV r^{2} values, and computed the average of those arbitrary distributions. We repeated this procedure for many (100, 1000, and 10000) iterations to compute a distribution of arbitrary SV and LV r^{2} values and consequently, the difference of those arbitrary distributions, which represents the null hypothesis. We treated the area under the null distribution corresponding to the difference between the true SV and LV r^{2} distributions as the pvalue. This procedure yielded identical results with 100, 1000, and 10000 permutations. It was performed for the model comparisons in Figure 3F, as well as for evaluation of the SV vs. LV models for the auditory (clicks) data.
To compare the different versions of the signal detection theorybased model (Figure 4), for example single accumulator version vs two accumulator version or two accumulator version vs. the two accumulator timevarying version, we used a bootstrapping approach. One hundred surrogate datasets were created by selecting behavioral trials at random with replacement from the original dataset. The signal detection theorybased model was then fit to each surrogate dataset, and the log likelihood for the bestfit parameters was recorded. This gave us a distribution of log likelihoods for each version of the model. We then performed the nonparametric permutation test described above on the distributions of log likelihoods to make pairwise comparisons between the bestfit model and all other models.
Acknowledgments
The authors would like to thank Klaus Osorio and Jovanna Teran for assistance with behavioral training and animal husbandry. Bingni Brunton contributed data that were collected for a previous paper (Brunton et al., 2013). Tim Hanks provided valuable comments on the manuscript and was very helpful in early stages of the analysis. Sam Lewallen and Mikio Aoi suggested useful analytic approaches. In addition the authors would like to thank all members of the Tank and Brody labs for useful discussions.
Appendix 1
Let
${N}_{R}=$ number of Right pulses, ${N}_{L}=$ number of Left pulses,
and let
where ${k}_{0}$ and $k$ are constants, and $N$ will represent number of pulses on one of the two sides.
Then define
where
Consider the case where $k{N}_{L}\gg {k}_{0}$ and ${N}_{R}=0$ (the easiest “go Left” trials). Under these conditions ${k}_{0}$ is negligible, and we obtain
A similar expression holds for the easiest “go Right” trials. Equation (5) corresponds to the “lapse rate” according to the model: that is, the probability of making an error even in the easiest trials. Notice (a) that the lapse rate is nonzero even though there is no explicit lapse rate parameter and (b) ${k}_{0}$ does not enter into the expression for the lapse rate.
Now consider the case where the number of pulses is very low (e.g., ${N}_{R}=0$ and ${N}_{L}=1$). The larger that ${k}_{0}$ is, the closer that P (choice = RightN_{R}, N_{L}) ≈ 0.5, in other words, the closer that performance will approach chance. Consequently, for large ${k}_{0}$, performance starts near chance for very short duration trials (since these will have low numbers of clicks). Performance will grow for longer duration trials, and will reach a maximum fraction correct of 1 – erf (–1/k) for the easiest, longest, trials.
References

Functional dissection of circuitry in a neural integratorNature Neuroscience 10:494–504.https://doi.org/10.1038/nn1877

Three factors limiting the reliable detection of light by retinal ganglion cells of the catThe Journal of Physiology 200:1–24.https://doi.org/10.1113/jphysiol.1969.sp008679

The analysis of visual motion: a comparison of neuronal and psychophysical performanceThe Journal of Neuroscience 12:4745–4765.

The detection of visual contrast in the behaving mouseThe Journal of Neuroscience 31:11351–11361.https://doi.org/10.1523/JNEUROSCI.668910.2011

Estimates of the contribution of single neurons to perception depend on timescale and noise correlationThe Journal of Neuroscience 29:6635–6648.https://doi.org/10.1523/JNEUROSCI.517908.2009

Linearnonlinearpoisson models of primate choice dynamicsJournal of the Experimental Analysis of Behavior 84:581–617.https://doi.org/10.1901/jeab.2005.2305

Development of elementary numerical abilities: a neuronal modelJournal of Cognitive Neuroscience 5:390–407.https://doi.org/10.1162/jocn.1993.5.4.390

Subtracting pigeons: logarithmic or linear?Psychological Science 12:244–246.https://doi.org/10.1111/14679280.00343

Temporal accumulation of visual evidence in the rat

Nonverbal numerical cognition: from reals to integersTrends in Cognitive Sciences 4:59–65.https://doi.org/10.1016/S13646613(99)014242

Scalar expectancy theory and weber's law in animal timingPsychological Review 84:279–325.https://doi.org/10.1037/0033295X.84.3.279

[means of restricting the movements of conscious rats in neurophysiologic experiments]Zhurnal Vyssheĭ Nervnoĭ Deiatelnosti Imeni I P Pavlova 30:1087–1089.

Responses of neurons of primary visual cortex of awake unrestrained rats to visual stimuliNeuroscience and Behavioral Physiology 15:379–386.https://doi.org/10.1007/BF01184022

The relative influences of priors and sensory evidence on an oculomotor decision variable during perceptual learningJournal of Neurophysiology 100:2653–2668.https://doi.org/10.1152/jn.90629.2008

Relative and absolute strength of response as a function of frequency of reinforcementJournal of the Experimental Analysis of Behavior 4:267–272.https://doi.org/10.1901/jeab.1961.4267

Multisensory decisions provide support for probabilistic number representationsJournal of Neurophysiology 113:3490–3498.https://doi.org/10.1152/jn.00787.2014

Integration of direction cues is invariant to the temporal gap between themThe Journal of Neuroscience 33:16483–16489.https://doi.org/10.1523/JNEUROSCI.209413.2013

Bounded integration in parietal cortex underlies decisions even when viewing duration is dictated by the environmentThe Journal of Neuroscience 28:3017–3029.https://doi.org/10.1523/JNEUROSCI.476107.2008

Active sensation: insights from the rodent vibrissa sensorimotor systemCurrent Opinion in Neurobiology 16:435–444.https://doi.org/10.1016/j.conb.2006.06.009

A role for neural integrators in perceptual decision makingCerebral Cortex 13:1257–1269.https://doi.org/10.1093/cercor/bhg097

A mode control model of counting and timing processesJournal of Experimental Psychology 9:320–334.https://doi.org/10.1037/00977403.9.3.320

Representation of number in the brainAnnual Review of Neuroscience 32:185–208.https://doi.org/10.1146/annurev.neuro.051508.135550

Multisensory decisionmaking in rats and humansThe Journal of Neuroscience 32:3726–3735.https://doi.org/10.1523/JNEUROSCI.499811.2012

Dual diffusion model for singlecell recording data from the superior colliculus in a brightnessdiscrimination taskJournal of Neurophysiology 97:1756–1774.https://doi.org/10.1152/jn.00393.2006

Simultaneous cellularresolution optical perturbation and imaging of place cell firing fieldsNature Neuroscience 17:1816–1824.https://doi.org/10.1038/nn.3866

A computational analysis of the relationship between neuronal and behavioral responses to visual motionThe Journal of Neuroscience 16:1486–1510.

Psychology and neurobiology of simple decisionsTrends in Neurosciences 27:161–168.https://doi.org/10.1016/j.tins.2004.01.006

The time course of perceptual choice: the leaky, competing accumulator modelPsychological Review 108:550–592.https://doi.org/10.1037/0033295X.108.3.550

Subitizing: the role of figural patterns in the development of numerical conceptsArchives De Psychologie 50:191–218.
Decision letter

Naoshige UchidaReviewing Editor; Harvard University, United States
In the interests of transparency, eLife includes the editorial decision letter and accompanying author responses. A lightly edited version of the letter sent to the authors after peer review is shown, indicating the most substantive concerns; minor comments are not usually included.
Thank you for submitting your work entitled "Sources of noise during accumulation of evidence in unrestrained and voluntarily headrestrained rats" for peer review at eLife. Your submission has been favorably evaluated by Eve Marder (Senior Editor), Maoshige Uchida (Reviewing Editor), and three reviewers, one of whom, Patrick Simen, has agreed to share his identity.
The reviewers have discussed the reviews with one another and the Reviewing Editor has drafted this decision to help you prepare a revised submission.
All the reviewers thought that the experiments are well done, the analyses are solid (but see below), and the conclusions are very interesting. In particular, the reviewers thought that the use of a decision paradigm in which the exact information provided to an animal on each trial can be precisely controlled is a powerful means to study the mechanisms underlying decision making. This property indeed allowed the authors to analyze the decisionmaking process in a greater detail. The authors conclude that the standard deviation (SD) of the internal estimates grows linearly with the number of flashes (i.e. features Weber scaling), which violates the linear scaling in variance (rather than SD) commonly assumed by evidence accumulation models (including the one used in Brunton et al. [2013]). On the other hand, the reviewers raised some important issues that the authors should address before publication in eLife.
Essential revisions:
1) Model incompatibility. Some of the important conclusions are based on the fit parameters obtained using a drift diffusion (DDF) model (Brunton et al., 2013). However, the authors' analysis using the signal detection theory shows that some of the important assumptions in DDF models are wrong. How can the authors justify the conclusions obtained by the DDF model? The authors should analyze the data further to address the issue of whether and how the incompatibility between these models affects the validity of their conclusions.
2) The authors should perform model comparisons more quantitatively e.g. using AIC/BIC. The reviewers raised the issue that unlike what is stated in the text, the models might have different numbers of parameters. Please clarify. Also, model comparisons should be done at the level of individual animals as well as populations.
3) The reviewers thought that some of the assumptions such as zero drift during interpulse intervals are not justified.
4) The authors should show more results regarding the variability between animals.
5) How bootstrap methods were applied needs to be explained more explicitly. If the analyses were run such that the number of resampling affects the significance level, then the statistical results (pvalue) do not appear to be valid. Please clarify.
6) The model, as well as the data, seems to suggest that scaling of the numbers of pulses by the same factor will lead to identical discriminability. This property seems important. Please clarify whether this is true in the model as well as the data.
https://doi.org/10.7554/eLife.11308.025Author response
1) Model incompatibility. Some of the important conclusions are based on the fit parameters obtained using a drift diffusion (DDF) model (Brunton et al., 2013). However, the authors' analysis using the signal detection theory shows that some of the important assumptions in DDF models are wrong. How can the authors justify the conclusions obtained by the DDF model? The authors should analyze the data further to address the issue of whether and how the incompatibility between these models affects the validity of their conclusions.
We thank the reviewers for helping us greatly clarify this aspect of the original manuscript. We have revised the current manuscript to address this issue using the following approaches:
A) The critical features used in our signal detection theory (SDT) model are that diffusion noise in the interpulse interval is nearzero, and that pulses from throughout the entire stimulus interval contribute roughly equally to the animal’s decision. We have performed additional modelfree approaches to specifically support these claims (Figure 2; Figure 2—figure supplement 2, Figure 2—figure supplement 3) without reference to the Brunton et al. model, and we have rewritten the text to make it clear that our conclusions are based primarily on modelfree analyses, and do not require the DDF model.
B) We have rewritten the text to place less emphasis on the DDF model from Brunton et al., and as above, to make it clear that the SDT model and its conclusions do not depend on the Brunton et al. model. As the reviewers point out, one of the main messages of our paper is that the Brunton et al. model made some unjustified assumptions, so it makes no sense to depend on it.
We nevertheless still include, in the supplementary material, the results of applying the Brunton et al. model to our visual task data, so as to facilitate direct comparison between analyses of our task and the related previous auditory pulses task of Brunton et al. (It is interesting to note that the Brunton et al. fits for our visual task are qualitatively consistent with the fits to the previous auditory task – i.e., that model alone is not sufficient to distinguish qualitative differences between the two tasks.) Similarly, we report the results of applying our SDT model to both the visual task data and the auditory task data.
It is also interesting to note that, although imperfect, the interpretations that can be drawn from the Brunton et al. DDF model are consistent with the modelfree analyses on which the paper now rests. We hope that readers interested in drift diffusionlike models of decision making will find the aspects of the DDF model that are robust across studies – as well as its shortcomings – informative.
2) The authors should perform model comparisons more quantitatively e.g. using AIC/BIC. The reviewers raised the issue that unlike what is stated in the text, the models might have different numbers of parameters. Please clarify. Also, model comparisons should be done at the level of individual animals as well as populations.
We thank the reviewers for pointing this out: they are quite right, in our original submission we did indeed compare models that had different numbers of parameters but failed to use metrics that penalize model complexity, such as AIC. The question of which specific penalization to use for comparing nonnested models with different parameters (AIC, BIC, DIC, “empirical bayes” etc) is a subject of ongoing debate within the machine learning literature. The reviewers’ comment prompted us to perform a new analysis, which we feel is far simpler and does not suffer from the complications of statistical model comparison for nonnested models. We simplified our models so that all of the 8 model variants that we compare have exactly the same number of parameters (only 2), with similar structures for the two parameters across all the 8 models. This approach made comparison across models much more straightforward, and allowed us to compare log likelihoods directly. The results of this new comparison are described in our Results section and in our revised Figure 4. We also include a new figure (Figure 4—figure supplement 1) that compares the different models for each individual animal.
3) The reviewers thought that some of the assumptions such as zero drift during interpulse intervals are not justified.
As we understood it, the reviewers’ suggestion here is that a flash might create a longlasting input to the accumulator (i.e., nonzero drift during interpulse intervals), whereas we assumed zero drift during interpulse intervals. It is worth noting that if a pulse created a shortlasting non zero drift (e.g., on the order of tens of milliseconds), our results would remain essentially the same, because all analyses and results depend on the integrated effect of each pulse. The important question is thus with respect to longlasting nonzero drift.
The longer a nonzero postpulse drift, the greater its cumulative impact on the accumulator. Thus we would expect decisions to be biased toward the side with the greatest cumulative postpulse time. To examine whether our assumption of zero drift during interpulse intervals was reasonable, we implemented a regression model with two parameters: cumulative postpulse interval duration and difference in flash number (plus a third parameters, for a constant term). Longlasting postpulse drift would produce a significant positive regression weight for the cumulative postpulse duration. Instead, the model revealed that the rats’ choices were better explained by differences in flash number on each trial. More specifically, the regression coefficient for the difference in flash number was. 25 +/ .0038, compared to .06 +/ .0036 for cumulative postpulse interval duration (and. 02 +/ .0012 for the bias/constant term). We concluded that nonzero drift during the interpulse intervals did not appreciably contribute to the rats’ behavior.
4) The authors should show more results regarding the variability between animals.
We thank the authors for this suggestion. We have included twelve additional supplementary figures to address this issue. Figure 1—figure supplement 2 illustrates the psychometric performance of each individual rat. Figure 2—figure supplement 1 plots the bestfit model parameters to the drift diffusionlike model from Brunton et al. (2013) for each rat. Figure 2—figure supplement 2 and Figure 2—figure supplement 3 illustrate the effects of flash number and trial duration on behavioral performance for each individual rat. Figure 2—figure supplement 4 illustrates the psychophysical kernel (i.e. reverse correlation) of each individual rat. Figure 3—figure supplement 1 plots the bestfit model parameters to the signaldetection theory model for each rat. Figure 3—figure supplement 2 compares the signal detection theory model prediction to the data from each rat. We have also included additional analyses of each rat that performed the auditory (clicks) version of this task. Figure 3—figure supplement 4 plots the fits of the SDT model for each rat from the clicks task, and Figure 3—figure supplement 5 plots the result of a permutation test for each clicks rat comparing the goodnessoffit of scalar variability versus linear variance. Figure 3—figure supplement 6 and Figure 3—figure supplement 7 plot the chronometric curve of each rat, as well as the signaldetection theory model prediction. Figure 4—figure supplement 1 compares model performance of all eight models described in Figure 4 for each rat in the visual version of the task. We have referenced these analyses in the main text and included appropriate supplementary figure legends.
5) How bootstrap methods were applied needs to be explained more explicitly. If the analyses were run such that the number of resampling affects the significance level, then the statistical results (pvalue) do not appear to be valid. Please clarify.
We thank the reviewers for pointing this out. We have recomputed significance using a nonparametric permutation test, so that the pvalue is not affected by amount of resampling, which we confirmed by observing that the procedure yielded the same results when performed with 100, or 1000, or 10,000 permutations.
In more detail, when comparing the distribution of goodness of fit values (R2) for the scalar variability (SV) and linear variance (LV) models in Figure 3F, the null hypothesis is that the R2 values for each distribution derive from a common distribution. [To clarify: the distribution of bootstrapped R2 values were obtained by resampling from the behavioral trials with replacement, performing the fitting procedure, and computing the R2 value, on 1000 iterations.] To test the null hypothesis, we combined the bootstrapped R2 values from the SV and LV models into a single distribution, and from that combined distribution, created two arbitrary distributions of fake SV and LV R2 values, and computed the average of those arbitrary distributions. We repeated this procedure for many iterations to compute a distribution of arbitrary SV and LV R2 values and consequently, the difference of those arbitrary distributions, which represents the null hypothesis. The area of the null distribution corresponding to the difference between the true SV and LV R2 distributions is the pvalue. As pointed out above, the procedure yielded identical results with 100, 1000, and 10000 permutations. It was performed for the model comparisons in Figure 3F, as well as for evaluation of the SV vs. LV models for the auditory (clicks) data. Additionally, to evaluate the models in Figure 4, this permutation test was performed on the distributions of bootstrapped likelihoods for each model. There is a new section of the Methods section that describes the bootstrapping and permutation procedures in full (see Methods: Model comparison).
6) The model, as well as the data, seems to suggest that scaling of the numbers of pulses by the same factor will lead to identical discriminability. This property seems important. Please clarify whether this is true in the model as well as the data.
We thank the reviewers for suggesting this very interesting analysis: they are indeed correct that it is an important property, and it led us to some further analyses that, among other things, helped to better understand the comparison between our visual task and the auditory task of Brunton et al.
First, we have included a new supplementary figure (Figure 3—figure supplement 3) confirming that, as predicted by the reviewers, scaling the number of pulses on both sides by the same factor leads to nearly identical performance. This is true in both the data and in the signaldetection theory model predictions. (The signal detection theory model does predict some deviations from scalar variability, at very low numbers of flashes. These are also observed in the data, as predicted, and are also now described in the main text).
Second, we observed that perfect scalar variability would imply that for the auditory task of Brunton et al., performance should not increase for longer stimulus durations; yet in the data itself, performance does clearly increase. When fitting the auditory data with the SDT model, we found that a linear relationship between noise standard deviation and number of pulses still applied (as in scalar variability), but with a significant nonzero offset k0, i.e., σN = k0 + kN. This nonzero k0 can account for the observed improvement in performance over time in the auditory data, and contrasts with data from the visual task, for which k0 ≈ 0 (compare Figure 3—figure supplement 6 with Figure 3—figure supplement 7).
Both of these are important further analyses that followed from the reviewers’ observations and that we think readers will find interesting. In particular, these analyses allowed us to strengthen comparisons between the current visual data and the previous auditory data, and they establish that the linear relationship σN = k0 + kN is not specific to the visual sensory modality.
The new main text discussing these points can be found in the section Results: “Subjects’ estimates of flash number exhibit scalar variability”.
https://doi.org/10.7554/eLife.11308.026Article and author information
Author details
Funding
National Institutes of Health (5F32NS78913)
 Ben B Scott
Helen Hay Whitney Foundation (Postdoctoral fellow)
 Christine M Constantinople
National Institutes of Health (R21NS082956)
 David W Tank
 Carlos D Brody
National Institutes of Health (U01NS090541)
 David W Tank
 Carlos D Brody
Howard Hughes Medical Institute (Investigator)
 Carlos D Brody
The funders had no role in study design, data collection and interpretation, or the decision to submit the work for publication.
Ethics
Animal experimentation: Animal use procedures were approved by the Princeton University Institutional Animal Care and Use Committee (IACUC) (Protocol #1837 and #1853). These procedures were carried out in accordance with the recommendations in the Guide for the Care and Use of Laboratory Animals of the National Institutes of Health.
Reviewing Editor
 Naoshige Uchida, Harvard University, United States
Publication history
 Received: September 2, 2015
 Accepted: December 15, 2015
 Accepted Manuscript published: December 17, 2015 (version 1)
 Version of Record published: February 1, 2016 (version 2)
Copyright
© 2015, Scott et al.
This article is distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use and redistribution provided that the original author and source are credited.
Metrics

 3,686
 Page views

 1,033
 Downloads

 35
 Citations
Article citation count generated by polling the highest count across the following sources: Scopus, Crossref, PubMed Central.