TY - JOUR AU - Shi, Zhuanghua AB - Introduction Selecting the most relevant, and deprioritizing irrelevant, visual information is important in many everyday tasks. These priorities need to be flexibly updated based on current goals as well as on prediction about what visual dimensions and features will need to be attended to and processed in pursuing these goals. Visual search tasks provide a rich source of evidence for studying the dynamics underlying this updating: Apart from more general goal-based, top-down adjustments of the task set, the system settings are constantly adapted in response to the stimuli encountered on sequential search trials. These adaptations give rise to inter-trial, or selection, history effects, where performance on the task improves when some critical stimulus property is repeated across trials and impaired if it changes (e.g. [1]). Such inter-trial effects have been documented consistently for cross-trial repetitions/switches of specific target-defining features, such as color, spatial frequency [2,3], and size [4], for whole target-defining dimensions (e.g. [5,6]), and for target position [7,8]. Importantly, models of the updating of attentional control settings make predictions about these inter-trial effects that can be examined against the data (e.g. [9,10]), thus helping us understand the underlying computational principles. Accordingly, the goal of the present study was to apply mathematical modeling–specifically, evidence accumulation models to RT data to a ‘classical’ visual singleton, or ‘pop-out’, search task (adapted from [3,8]), employing a novel approach, namely: taking into account observers’ natural perceptual uncertainty in performing a given task and, associated with this, the possibility of adaptive trial-by-trial updating of the model parameters, in order to characterize the perceptual and cognitive mechanism/s of multiple types of (non-/spatial) inter-trial ‘priming’ in this task. Paradigms examining for inter-trial effects differ with regard to how long-lasting the respective (feature-, dimension-, position-specific) inter-trial effects are. One paradigm that has been documented to produce inter-trial effects of longer durations is known as ‘priming of pop-out’ (PoP; e.g. [2,3,8,11]). In a previous modeling study [12], we used a paradigm in which a pop-out target was presented in a dense search array (consisting of 39 closely spaced items) and the features of the (homogeneous) distractor items never changed across trials. In that study, described in more detail below, we only found significant inter-trial effects from a single trial back. By comparison, the PoP paradigm produces inter-trial effects that can be traced back for approximately five to eight trials for (repetitions/switches of) the target-defining feature [3] as well as for the target position [8,13]. PoP studies typically use sparse search displays (e.g., with three, widely spaced items only) and, importantly, random swapping across trials of the search-critical target and distractor features (e.g., the target may be red amongst green distractors on one trial, and green amongst red distractors on the next trial). Of note, feature-based inter-trial effects manifest specifically with sparse displays and target-distractor color swapping, but not or to a lesser extent with dense displays or constant distractor color [13,14]. That is, in PoP paradigms, guidance based on the computation of local feature-contrast, the major determinant of bottom-up saliency (e.g., see [15]), is relatively ineffective (because there is little local feature contrast), and in contrast to the name ‘priming of pop-out’, the target actually fails to be the first item to draw attention in a significant proportion of trials (e.g., Rangelov et al. [16] estimate this proportion to be of the order of 20% to 70%, consistent with eye-movement evidence such as from Becker [17]). Given that (bottom-up) saliency coding is a relatively non-reliable guide to finding the target, and that the target is consistently defined in the color dimension, the search-guidance system comes to rely on other types of information to optimize target selection under PoP conditions: in particular, top-down–color-feature-based–guidance processes, along with reliance on positional information (also evidenced by eye movements often returning, on trial n, to the target location on trial n-1). The roles of these, while traceable in dimension-based paradigms (feature-based especially with color-defined targets; e.g., [5]), become much more prominent under PoP conditions, and empirically, they exhibit a more persistent (n-back) effect than dimension-based priming effects. Thus, the underlying mechanisms, that is, the neural machinery that is primed, are likely to be different. Accordingly, attempting to model these (feature- and position-based inter-trial) effects using our modeling framework provides a theoretically interesting extension to our previous study. Further, feature-based inter-trial effects in PoP paradigms are typically independent of, or additive to, positional intertrial effects [8,18]. Some PoP studies also found inter-trial effects for the response-critical target feature, such as whether a section (“notch”) of the (color pop-out) target is cut off at the top or the bottom [19–21], while other studies failed to find such effects [3,22]. Interestingly, Gokce et al. [19] found the effect of response repetition to be dependent on position repetition: response repetition, from one trial back, expedited RTs significantly only when the target position was also repeated (similar results have also been found by [23–25]). One reason why some studies failed to find significant inter-trial effects for the response feature may be that target position repetitions were relatively rare. For instance, in the studies by Maljkovic et al. [3,22], the target occurred at one of twelve positions, randomly chosen from trial to trial; that is, the target position would have repeated on only some 8% of the trials, and there would have been a response repetition on only half these trials, making it hard to resolve the interaction. While there have been many previous studies on inter-trial effects in visual search, most of these have focused on mean response times (RTs) (and error rates). However, more information can be obtained from the shape of the RT distribution. For a variety of different tasks, the shapes of RT distribution shapes are well predicted by evidence accumulation models [26,27], with the ‘drift-diffusion model’ (DDM) being one particularly influential model of this type [28–30]. The DDM assumes that a decision between two different perceptual hypotheses is made by accumulating multiple pieces of evidence over time, summing the logarithm of the likelihood ratio, under the two different hypothesis, for these pieces of evidence, and making a decision when the sum reaches a threshold (in either direction; see Fig 1). This model has four important parameters: the drift rate, that is, a tendency to, on average, drift towards one or the other boundary (representing the average strength of the pieces of noisy evidence in favor of the respective hypothesis); the separation between the decision boundaries (boundary separation); a starting point (representing any initial bias, e.g., from a-priori priors or selection history); and a non-decision time (NDT, representing time spent on processes that are not part of the perceptual decision making as such, such as preparation and execution of the motor response). Another evidence accumulation model is the ‘Linear Approach to Threshold with Ergodic Rate’ (LATER) model [31,32]. This model also assumes that evidence accumulates until it reaches a threshold, but, unlike the DDM, it assumes that during a single perceptual decision, evidence accumulates at a constant rate, with this rate varying randomly across search episodes (see Fig 1). Download: PPT PowerPoint slide PNG larger image TIFF original image Fig 1. Evidence accumulation models. The DDM assumes that evidence accumulates, from the starting point (S0), through random diffusion in combination with drift at a constant rate r until a boundary (i.e., threshold, θ) is reached (illustrated in blue). The Linear Approach to Threshold with Ergodic Rate (LATER) model (illustrated in red) makes the same assumptions, except that there is no random diffusion, but instead the rate r varies across trials (so as to explain trial-to-trial variability in RTs). In addition, a non-decision time (NDT) τ is added to the boundary crossing time on each trial, to capture time spent on everything else than the perceptual decision (e.g., the time to prepare and execute a selected motor response). https://doi.org/10.1371/journal.pcbi.1009332.g001 In a previous study [12], we fit the DDM and LATER models to RT distributions from three different pop-out visual search experiments, in order to examine which model parameters were involved in different types of inter-trial effects occurring in the respective paradigms. In two of these experiments, we used a block-wise frequency manipulation of the response-critical feature (RCF) of the search displays. In one of these experiments, the response was made based on whether a target was present or absent, and in the other based on whether the target was defined in the color or the orientation dimension. In order to model inter-trial effects, we assumed that the starting point and the drift rate of the evidence accumulation model each could change after each trial based on two stimulus properties on that trial: the RCF and the target-defining dimension (TDD). We considered a number of different updating rules, different plausible rules of how these parameters could change, such that the RCF and TDD each could influence either the starting point or the drift rate (or neither) of the evidence accumulation process, and compared these in a factorial model comparison. This comparison revealed that the best combination of updating rules for the RCF and the TDD (in terms of the Akaike Information Criterion, AIC) involved updating the starting point based on the RCF, using a form of Bayesian updating consistent with an interpretation of the starting point as the logarithm of the prior odds, in this case the prior odds of the target being present versus absent or of being defined by color versus orientation; and updating the drift rate based on the TDD, using a ‘weighted-rate’ updating rule consistent with the ‘dimension-weighting account’ [5,6]. This model captured both the effects of the probability manipulation and the inter-trial effects based on the RCF and the TDD quite well. However, the inter-trial effects were of relatively short duration: significant inter-trial effects were resolvable only from a single trial back, as is generally the case for ‘dimension-weighting’ studies [14]. Accordingly, we could not draw any strong conclusions about how well our model would capture the dynamics of longer-lasting inter-trial effects described in the literature [33,34] and the decay of the memory traces underlying these inter-trial effects over time. In the present study, we aimed to address this question by applying our modeling framework to the PoP paradigm which is known to produce longer lasting inter-trial effects. As elaborated above, each type of inter-trial effect–faster RTs with repetition, compared to change, of the target-defining feature, the target position, and the response-critical feature–could be mediated by different mechanisms, such as more efficient target processing or responding based on less evidence (i.e., a response bias). Depending on the underlying mechanism, one can make different predictions about how RT distributions will differ across inter-trial conditions, such as repetition versus switch of the target-defining color. In our modeling framework, such differences in RT distributions across different inter-trial conditions are predicted by changes in the evidence accumulation model parameters based on stimulus and selection history. Thus, by comparing different updating rules, which differ in terms of which model parameters change based on the history of each stimulus attribute, and finding the updating rule that best explains the RT distributions, we can make inferences about the likely underlying mechanism for each type of inter-trial effect. To this end, we applied this approach to the data set collected by Gokce et al. [19], which revealed all three types of inter-trial effects considered above: faster RTs for repetitions (vs. changes) of the target-defining feature (color: green vs. red), the target position, and the response-critical feature (the position of a “notch”, top vs. bottom, in the target item). Fig 2 illustrates the stimuli and inter-trial transitions used in Gokce et al. [19]. In that study, participants were required to find a singleton target stimulus defined by an odd-one-out color (the only green item amongst red items, or the only red item amongst green items) and respond according to whether the target had a cut-off section (notch) at the top or the bottom. The target and distractor stimuli were positioned on a virtual circle, forming either a regular square or diamond arrangement on a given trial. Across consecutive trials n-1 and n, the target/distractor color polarity could either repeat (e.g.: trial n-1: red target amongst green distractors → trial n: red target amongst green distractors) or switch (e.g., trial n-1: red target amongst green distractors → trial n: green target amongst red distractors), with equal probability. Independently of this, the target on a given trial (n) could appear at the same location as on the previous trial (n-1), henceforth referred to as a TT trial; it could appear at a previously empty, or neutral location, henceforth labelled TN trial; or it could occur at a previous distractor location, TD trial; again, all three positional inter-trial transitions (see Fig 2 for an illustration) were equally likely. Finally, the ‘orientation’ of the response-critical notch in the target item (top or bottom) could be either repeated or switched across consecutive trials, again with equal probability. Given this, the data allowed for the examination of target feature-, position-, and response-based inter-trial priming effects; note that with regard to positional intertrial priming, the design made it possible to dissociate facilitatory (comparison TT vs. TN) and inhibitory (TN vs. TD) priming with respect to the neutral (TN) condition. Download: PPT PowerPoint slide PNG larger image TIFF original image Fig 2. Illustration of the stimulus. Illustration of the search display and the positional inter-trial transitions. The circles were not part of the actual stimulus, but are only shown to illustrate how the items were positioned on a circle. On each trial, four out of eight possible locations were occupied by the search items, allowing the items to form either a square or a diamond arrangement. https://doi.org/10.1371/journal.pcbi.1009332.g002 In summary, by comparing different models of how RTs are affected by stimulus history, we aimed to address a number of questions. The primary aim was to examine which types of updating can better predict the temporal profile of n-back inter-trial effects depending on whether some target feature (i.e., the target color and/or the ‘orientation’ of the notch), or the position, was repeated or changed. In addition, we aimed to investigate the degree of spatial specificity of the inter-trial effects for the response-critical target feature and the target-defining feature (here color), by comparing updating rules with differing degrees of position specificity of the parameter updates. Finally, by comparing updating rules operating on different parameters of the evidence accumulation model, we also aimed to contribute evidence towards understanding the underlying mechanism for each type of inter-trial effect. Different updating rules implement different hypotheses about what type of memory is carried over across trials and causing the priming effects. This may be a stimulus-response (S-R) binding carried over from a previous trial, resulting in a bias towards repeating the same response or perceptual decision (e.g., about the target location), or it may be a shift of attentional weight towards the target location or feature that is carried over across trials. Thus, by finding the best updating rule for each type of inter-trial effect, we aimed to clarify both the nature of the memory responsible for the inter-trial effect, the spatial specificity of this memory, and the speed of memory decay. Results Behavioral results Fig 3 depicts the mean RTs as a function of the response-critical target notch position (i.e., effectively, the response: same vs. different relative to the preceding trial) for trials with a repetition versus a switch of the target-defining color (same vs. different) across consecutive trials, separately for the three inter-trial target location transitions (target at previous target location, TT vs. at previously empty neutral location, TN vs. at previous distractor location, TD). A repeated-measures ANOVA with color (repetition/switch), response (repetition/switch), and target position (TT, TN, TD) as factors revealed all main effects to be significant (response: F(1, 13) = 7.75, p = .015, = 0.37, BFincl > 1000; color: F(1, 13) = 160.9, p < .001, = 0.93, BFincl > 1000; position; F(1.2, 15.6) = 72.82, p < .001, = 0.85, BFincl > 1000, Huynh-Feldt corrected degrees of freedom). RTs were faster when the target-defining color repeated vs. changed (48-ms difference), and when the response-critical notch position repeated vs. changed (16-ms difference). And RTs were significantly faster when the target appeared at the same position compared to either a previous distractor location (TT vs. TD: 46-ms difference, Bonferroni-corrected t(13) = 9.28, pbonf < .001, BF10 > 1000) or a previously empty (neutral) location (TT vs. TN: 30-ms difference, t(13) = 7.15, pbonf < .001, BF10 > 1000); there was also a significant cost when the target appeared at a previous distractor position vs. a previously empty (neutral) position (TD vs. TN: –16-ms difference, t(13) = –9.48, pbonf < .001, BF10 > 1000). In addition, the interactions RCF × position (F(1.4, 19) = 23.3, p < .001, = 0.64, BFincl > 1000 and RCF × color (F(1,13) = 6.19, p = .027, = 0.32, BFincl = 0.56) were significant, although the RCF × color interaction was not supported by the Bayesian analysis. Repeated-measures ANOVAs conducted separately for each positional inter-trial transition condition with response (repetition/switch) and color (repetition/switch) as factors revealed that the effects involving response (position of the target notch) were significant only in the repeated target position (TT) condition (main effect of response: F(1,13) = 32.6, p < .001, = 0.72, BFincl > 1000, response × color interaction: F(1,13) = 6.0, p = .029, = 0.32, BFincl = 0.89), but not in the TN or TD conditions (all p > .3). Download: PPT PowerPoint slide PNG larger image TIFF original image Fig 3. Response times. Mean response times for repeated/non-repeated color, target notch ‘orientation’ (the response- critical feature), and for the different positional transitions: target at previous target position (TT), target at previously neutral (i.e., empty) position (TN), and target at previous distractor position (TD). Error bars show the 95% confidence intervals. RCF: response-critical feature. https://doi.org/10.1371/journal.pcbi.1009332.g003 Model comparison results In our modeling framework (see Fig 4), we treat each trial of the experiment as a perceptual decision, which is modeled as an evidence accumulation process, and we allow the parameters of that evidence accumulation process (i.e., the starting point, rate of evidence accumulation, and non-decision time) to change from trial to trial based on recent stimulus history. Download: PPT PowerPoint slide PNG larger image TIFF original image Fig 4. Illustration of the hierarchical modeling framework. Each hierarchical model consists of an evidence accumulation model (either the drift-diffusion model or the LATER model) and either no updating or one updating rule for each parameter of the evidence accumulation model (starting point S0, evidence accumulation rate r, and non-decision time τ). Each updating rule belongs to one of the four categories shown in the middle layer of the figure, and is applied to one of the three inter-trial updating variables shown in the blue box on the right side of the figure: the response-critical feature (RCF), the target-defining color, or the target position. For some of the updating rules based on the RCF or color, there are three different versions of the rule, differing in their degree of position-specificity (top level of the hierarchy). These rules could be fully position-independent (PI), fully position-dependent (PD), with a gradient-like dependence on the change of position (PG), or starting out fully position-dependent but then spreading (PS). See the main text for detailed descriptions. https://doi.org/10.1371/journal.pcbi.1009332.g004 In particular, we consider three aspects of stimulus history: the cross-trial history of the response-critical feature (RCF), of the target color, and of the target position. For each of these, we consider different updating rules, each implementing a possible way in which one model parameter could change based on this aspect of stimulus history (see detailed mathematical description in the section of “Models and updating rules”). The aspect of stimulus history that an updating rule is based on is referred to as the updating variable (UV) of that rule. Each of the updating rules belongs to one of four categories: Bayesian rules, step rules, weighted rules, and binary rules. Binary rules have two different values of the updated parameter, one on trials where the UV was repeated from the immediately preceding trial (n-1) and the other on trials where it changed. Binary rules were included for comparison, serving as a ‘baseline’ to assess how much better inter-trial effects could be explained when taking into account trial history further back than n-1; step rules, weighted rules, and Bayesian rules represent three different ways of doing this. Step rules assume that repetition effects were partially carried over to future trials. For example, if the evidence accumulation rate was faster because the target color had been repeated between trial n-1 and trial n, some of this repetition benefit would be carried over to trial n+1 (and a smaller proportion carried over to trial n+2 and so on). Weighted rules instead assume that each state of the UV had an associated weight which determined the value of the updated parameter on trials where that state of the UV occurred. After each trial, some of the weight was shifted to the state of the UV which occurred on that trial, in such a way that the total weight remained constant, as if a limited resource was being reallocated. There was also memory decay of previous weight reallocations. The Bayesian rules were applied specifically to the starting point parameter, and assumed that the relative frequencies of the different states of the UV were learned through Bayesian updating, with memory decay as in the dynamic belief model of Yu and Cohen [35]. These frequencies were assumed to define a prior for the evidence accumulation process on each trial, implemented by setting the starting point to the logarithm of the prior odds of the state of the UV on that trial. For some of the updating rules based on the RCF and the target color, we also compared four different versions of the rule, differing in their degree of position specificity as well as in whether memories learned at a particular position would spread over time. These rules could be fully position independent (PI), fully position-dependent (PD), with a gradient-like dependence on the change of position (PG), or initially fully position-dependent but with spreading over time (PS). The PI rule assumes that the influence from RCF or color on previous trials does not depend at all on whether the target position was the same or had changed, while the PD rule assumes no influence at all from a previous trial unless the same target position is repeated. By contrast, the rule of PG suggests a stronger influence from a previous trial the closer the target is to its previous position. Finally, the PS rule assumes that learning is initially fully position specific but later spreads (e.g. because the exact position in which a particular target appeared is gradually forgotten). For some of the position-based updating rules, there are three different versions, with and without inhibition of previous distractor locations which could be either fully matched with target location facilitation so that weight is only transferred from distractor to target locations, or involve separate processes of transferring weight to the target and away from distractors (including to and from empty locations). Further details about the updating rules and the model fitting are provided in the Methods and models section below. We compared different updating rules based on how well the model, when using this updating rule, predicts RTs on all trials and determined the best of these updating rules in terms of the Akaike Information Criterion (AIC; see Methods and models section for more details). The AIC is a measure of the quality of a model, which takes into account goodness of fit (as measured by the likelihood) and also penalizes models with more free parameters. Lower AIC values indicate better model performance. In total, we compared 12 different updating rules for updating based on the response-critical feature, eight for the target color and ten for the target position (taking into account also the two different evidence accumulation models DDM and LATER, there were thus a total of 2*12*8*10 = 1920 possible models). Figs 5, 6 and 7 show the mean relative Akaike Information Criteria (AIC) for each of the response feature-, target color-, and target position-related updating rules. For each of these three updating variables, the AIC for each updating rule was evaluated for a model which used the best updating rule, in terms of having the lowest associated AIC, for each of the other two updating variables (see Methods and models section for further details). For each individual participant and session, we subtracted the AIC of the overall winning model (based on all participants and sessions) from the AIC of every other model for that participant and session, and finally we averaged this relative AIC across all participants and sessions. This resulted in a relative AIC of zero for the winning model, while the relative AIC of other models indicate how much worse they are compared to the winning model. Download: PPT PowerPoint slide PNG larger image TIFF original image Fig 5. Model comparison for color-based updating. Mean relative AIC for all different color-based updating rules when using the best updating rules for response-critical feature (RCF) and position. The dashed vertical line separates rules that update based on color alone (left) and rules that also take the target position into account (right). The different background colors mark rules that update either the rate (orange) or the non-decision time (green). Error bars represent the standard error of the mean over participants and sessions. PI: position-independent; PD: full position-dependent; PG: a gradient-like dependence on the change of position; NDT: Non-decision time. https://doi.org/10.1371/journal.pcbi.1009332.g005 Download: PPT PowerPoint slide PNG larger image TIFF original image Fig 6. Model comparison for position-based updating. Mean relative AIC for all different position-based updating rules when using the best updating rules for response-critical feature (RCF) and color. The different background colors mark rules that update either the rate (orange) or the non-decision time (green). Error bars represent the standard error of the mean over participants and sessions. DI: distractor inhibition; NDT: Non-decision time. https://doi.org/10.1371/journal.pcbi.1009332.g006 Download: PPT PowerPoint slide PNG larger image TIFF original image Fig 7. Model comparison for response feature-based updating. Mean relative AIC for all different RCF-based updating rules when using the best updating rules for color and position. The dashed vertical line separates rules that update based on the RCF alone (left) and rules that also take the target position into account (right). The different background colors mark rules that update either the starting point (blue), the rate (orange) or the non-decision time (green). Error bars represent the standard error of the mean over participants and sessions. PI: position-independent; PD: fully position-dependent; PG: a gradient-like dependence on the change of position; NDT: Non-decision time; S0: Starting point. https://doi.org/10.1371/journal.pcbi.1009332.g007 Target color-based updating models. To find the best rule for target color-based updating, we compared the different target color-based updating rules in terms of the mean relative AIC (see Fig 5). The best rule was the position-independent weighted-rate rule (see updating rule 7 in the “Models and updating rules” subsection of the Methods and models section for details). This updating rule assumes that the evidence accumulation rate depends on how much of a limited “weight” resource, which is distributed across both of the possible target colors, is allocated to the target color. This weight is updated after each trial by shifting some weight to the target color on that trial, with partial “forgetting” of old updates (see Fig 8 for an example, Fig 9 for the predicted temporal profile with this updating rule, and Methods and models section for more details). Download: PPT PowerPoint slide PNG larger image TIFF original image Fig 8. Example of color-based updating. Example of the weight changes to the different colors predicted by the weighted-rate updating rule for the first eight trials of a typical participant (A), and the associated changes in the evidence accumulation rate on the same eight trials (B). The letters “T” and “D” denote the target color and, respectively, the distractor color on each trial. The dashed line marks the baseline evidence accumulation rate. The evidence accumulation rate on each trial was the baseline rate scaled by the weight associated with the target position on that trial. https://doi.org/10.1371/journal.pcbi.1009332.g008 Download: PPT PowerPoint slide PNG larger image TIFF original image Fig 9. Temporal profile of the color-based inter-trial effects. Mean normalized RT for repeated vs. switched target color on (current) trial n compared to (preceding) trial n-1 (lag 1), n-2 (lag 2) up to n-8 (lag 8). Normalized RTs were first averaged across the two sessions for each participant; the resulting (individual mean normalized) RTs were then used to compute the overall means and confidence intervals, across participants. Filled circles depict the behavioral data, lines the model predictions. Error bars represent 95% confidence intervals. https://doi.org/10.1371/journal.pcbi.1009332.g009 Target position-based updating models. Next, we compared the mean relative AIC of position-based updating rules (see Fig 6). The best rule for position-based updating was the weighted-rate with distractor inhibition rule (see updating rule 8 in the “Models and updating rules” subsection of the Methods and models section for details). This updating rule assumes that the evidence accumulation rate depends on how much of a limited weight resource, which is distributed across all the possible target positions, is allocated to the target position. This weight is updated after each trial by shifting weight to the target position on that trial from all other positions, both distractor positions and empty positions, and shifting weight away from the distractor positions on that trial to all other positions, both target positions and empty positions, with partial forgetting of old updates (see Fig 10 for an example and Fig 11 for the predicted temporal profile with this updating rule, and Methods and models section for more details). Download: PPT PowerPoint slide PNG larger image TIFF original image Fig 10. Example of position-based updating. Example of the weight changes to the different positions predicted by the “weighted rate with distractor inhibition” updating rule for the first eight trials of a typical participant (A), and the associated changes in the evidence accumulation rate on the same eight trials (B). The letters “T” and “D” denote the target position and, respectively, the positions of the three distractors on each trial. The dashed line marks the baseline evidence accumulation rate. The evidence accumulation rate on each trial was the baseline rate scaled by the weight associated with the target position on that trial. https://doi.org/10.1371/journal.pcbi.1009332.g010 Download: PPT PowerPoint slide PNG larger image TIFF original image Fig 11. Temporal profile of the position based inter-trial effects. Mean normalized RT for different positional inter-trial conditions, target in previous target condition (TT), target in previous distractor position (TD), target in previously neutral (unoccupied) position (TN), compared to trial n-1 (lag 1), n-2 (lag 2) up to n-8 (lag 8). Normalized RTs were first averaged across the two sessions for each participant; the resulting (individual mean normalized) RTs were then used to compute the overall means and confidence intervals, across participants. Filled circles show the behavioral data, while lines show model predictions. Error bars represent 95% confidence intervals. https://doi.org/10.1371/journal.pcbi.1009332.g011 RCF-based updating models. Finally, we compared the mean relative AIC of response-based updating rules (see Fig 7). The best rule was “Position Gradient (PG) Bayesian S0” (see Fig 12 for an example and Fig 13 for the predicted temporal profile and updating rule 4 in the “Models and updating rules” subsection of the Methods and models section for details). This updating rule learns a different starting point for each possible target position, updating the starting point primarily for the actual position where the target occurred, though with some updating carried over to other positions, with the magnitude of the update decreasing with distance (we refer to this a gradient-like dependence on the change of position). Interestingly, both the positionally non-specific version of the Bayesian starting point updating rule and the completely position-specific version (which updates only a single position with no effect on other positions) performed considerably worse. Download: PPT PowerPoint slide PNG larger image TIFF original image Fig 12. Example of response-based updating. Example of the starting point changes associated with the different positions predicted by the gradient-like dependence on the change of position (i.e., “PG Bayesian S0” updating rule) for the first eight trials of a typical participant (A), and the starting point for the target position on the same eight trials (B). The position of the triangles represent the target position on each trial, and the shape of the triangle indicates the response-critical feature (RCF, i.e., whether the notch was on the top or bottom of the target item). https://doi.org/10.1371/journal.pcbi.1009332.g012 Download: PPT PowerPoint slide PNG larger image TIFF original image Fig 13. Temporal profile of the response feature-based inter-trial effects. Mean normalized RT for repetition vs. switch of the response defining target feature (notch on top or bottom of the diamond shape), compared to trial n-1 (lag 1), n-2 (lag 2) up to n-8 (lag 8) for the different positional inter-trial conditions, target in previous target condition (TT), target in previous distractor position (TD), target in previously neutral (unoccupied) position (TN). Normalized RTs were first averaged across the two sessions for each participant; the resulting (individual mean normalized) RTs were then used to compute the overall means and confidence intervals, across participants. Filled circles show the behavioral data, while lines show model predictions. Error bars represent 95% confidence intervals. RCF: response-critical feature. https://doi.org/10.1371/journal.pcbi.1009332.g013 Cross-validation. In order to examine to what extent the better fit of the best updating rules compared to the ‘No-update’ rule was a result of overfitting, we performed cross-validation on the best model and the no-update model. Specifically, since each session consisted of eight blocks of trials, we performed an eight-fold cross-validation, that is: for each block, we evaluated the model’s prediction in that (‘test’) block after training the model on the remaining seven (‘training’) blocks (the stimulus sequence in the test block was still used for updating, so that the starting point and weight would start with the correct values at the beginning of the next block, but participants’ performance in that block was not used for optimizing the model parameters). Table 1 presents the average log-likelihood (logarithm of the likelihood) for the best model as well as the model with no updating, each evaluated on the training set (averaging the log-likelihood in each block across the seven folds in which that block was included in the training set) as well as on the test set. The average log-likelihood was somewhat worse when evaluated on the test set, indicative of some degree of overfitting, and this difference was larger for the best model compared to the no-update model, suggesting there was some overfitting of the updating rules, in addition to the evidence accumulation model. Importantly, however, even when evaluated on the test set, the log-likelihood was substantially better for the best model compared to the no-update model, indicating that the updating rules did capture patterns in the data that generalize across blocks. This conclusion is further supported by the cross-validated out-of-fold predictions for the temporal profiles of the inter-trial effects presented in S6 Appendix, which fit the pattern in the data nearly as well as the corresponding predictions without cross-validation (depicted in Figs 9, 11 and 13 below). Download: PPT PowerPoint slide PNG larger image TIFF original image Table 1. Average log-likelihood, across participants and sessions, for the best model and the no-update model, evaluated on the training set as well as on the test set, and the difference between the test and training set log-likelihood (right column). The bottom column shows the difference in log-likelihood between the best model and the no-update model. https://doi.org/10.1371/journal.pcbi.1009332.t001 Temporal profiles of the inter-trial effects and model predictions Having seen which updating rule best accounts for each type of inter-trial effect, in this section we will explore what prediction those updating rules make for the temporal profile of each inter-trial effect and how well this matches the temporal profiles in the experimental data. In the next three subsections, we explore how well the model framework predicts the three different types of inter-trial effect: (i) inter-trial effects related to repetition/switch of the target-defining color, more precisely: repetition of both the target and distractor colors versus swapping of the target and distractor colors; (ii) inter-trial effects related to the target position, for which there are three different transition conditions: target at previous target position, target at previously unoccupied (neutral) position, and target at previous distractor position; (iii) inter-trial effects related to the response-critical feature, the position (top or bottom) of the notch in the target item. All figures in the next three sections show normalized RTs, where the overall mean RT has been subtracted from the mean in each condition, for each participant and session (i.e., a positive normalized RT means that the response time is slower than overall average in a particular condition, while a negative normalized RT means it is faster than overall average). To make the model predictions directly comparable to the behavioral data, the mean RTs (from the behavioral data) were also subtracted from the model predictions. All model predictions are based on the best model, that is, using the factor level that resulted in the lowest AIC for each of the factors (see the model comparison section above); accordingly, the relevant figures below (Figs 9, 11 and 13) are all based on the exact same model (see S2 Appendix for the parameters of the model fits), applying all of the winning updating rules (even though each of these figures only illustrates the effects of one, or two in the case of Fig 13, of these rules). Color-based inter-trial effects. Fig 8 illustrates the predictions of the best updating rule for color-based inter-trial effects, that is, the “weighted-rate” rule, for the first eight trials of a typical participant. The two colors start out receiving equal weight (each with a weight of 1 so that the rate for each color is the baseline rate), and after each trial some weight is shifted to the target color on that trial, while additionally there is partial forgetting so that the weights are partially reset to their starting values (see modeling part of Methods and models section for details). The figure illustrates how repetition benefits arise in the model: On trials 2, 3, and 5, the same target color is repeated as on the previous trial; on these trials, the weight associated with the target color is larger than one because some weight has just been shifted to that color; on trials 4 and 6–8, by contrast, the target color has changed from the previous trial, resulting in weight of less than one associated with the target color because some weight was just shifted away from that color. The figure also illustrates influences from more than one trial back: the weight associated with the target color is larger on trial 3 compared to trial 2, because this is the second repetition in a row and the weight shifts accumulate; by contrast, the weight is considerably lower on trial 5 because this color repetition was preceded by a sequence of three trials with the other target color. The color-based inter-trial effects were predicted well by this updating rule (shown in Fig 9). Response times are faster (by 56 ms, on average) when the color from trial n-1 is repeated on trial n. The same pattern is found for the repetition versus switch of the color from more than one trial back, although the magnitude of the effect decreases with increasing lag. This pattern is captured well by the model, although the model somewhat underestimates the effect for longer lags (i.e., for influences from four or more trials back), while it, if anything, slightly overestimates the effect at the shortest lags (influences from one to three trials back). Position-based inter-trial effects. The best updating rule for target position-based inter-trial effects (illustrated in Fig 10), that is, the “weighted rate with distractor inhibition” rule, followed a similar logic to that for the color-based effects. All eight positions start with equal weight (each with a weight of 1 so that the rate for each position is the baseline rate), and after each trial some weight is shifted to the target position on that trial from all other positions (i.e., both distractor positions and unoccupied positions), and away from the distractor positions on that trial to all other positions (i.e., both the target position and unoccupied positions; see modeling part of Methods and models section for details). This results in increased weight associated with the previous target position and decreased weight associated with the previous distractor positions, while the weight associated with the neutral positions does not change much because they both lose weight to the target position and receive weight from the distractor positions. This explains why the model predicts facilitation of response times for “target at previous target position” transitions (TT) and a cost for “target to previous distractor position” transitions (TD) compared to when the target moves to a previously neutral (empty) position (TN). The figure also illustrates influences from more than one trial back: both trial 4 and trial 7 are TD transitions, but the weight associated with the target position is much lower on trial 7, because the distractors had appeared at that position an additional two times in the recent trial history. The model also predicts well for the position-based inter-trial effects (Fig 11). Compared to when the target on trial n appeared at a position which was neutral (unoccupied) on trial n-1 (TN), RTs were faster (by 36 ms, on average) when the target appeared at the previous target position (TT), and slower (by 17 ms) when the target appeared at a previous distractor position. The same pattern of positional inter-trial effects is seen for transitions from more than one trial back, though the size of the effect decreased with increasing lag. The model predicts this pattern, although it predicts a slightly larger target-target transition benefit relative to the distractor-target transition cost compared to the behavioral data. Response-based inter-trial effects. The best updating rule for the response-based inter-trial effects: “PG Bayesian S0” (illustrated in Fig 12) assumed that participants learned a separate response bias (starting point of the evidence accumulation model) for each position, based on the frequency with which each response has recently been associated with that position, but that updating of the bias in the current target position on any trial partially carried over to other positions in a gradient-like way. In the example shown in Fig 12, the starting point is biased towards the boundary associated with a “notch on top” response after the first trial, where the strongest bias is associated with the position where the target occurred but with some bias also for other positions. On the second trial, the target position was repeated but the notch position changed–so the bias learned from the first trial is nearly cancelled. However, because of the memory decay, the new notch position had a somewhat stronger influence than the old one; so, instead of complete cancellation, the result was a small bias in the other direction. The next five trials all had the notch on the top, but with the target occuring in different positions, resulting in a build-up of a bias towards that response associated with all positions. Because inter-trial effects for the response primarily occurred when the target position was repeated (see Fig 3), the RCF-based inter-trial effects are shown separately for the different positional inter-trial transition conditions (Fig 13). The model predicts well in the repeated target position condition (TT). However, for the other conditions, the model slightly overestimates the inter-trial effects from one trial back and somewhat underestimates those for longer lags. This happens because in the behavioral data, there are virtually no effects of repetition/switch of the RCF from a single trial back when the target position does not repeat, while there are some such effects from two and more trials back. Behavioral results Fig 3 depicts the mean RTs as a function of the response-critical target notch position (i.e., effectively, the response: same vs. different relative to the preceding trial) for trials with a repetition versus a switch of the target-defining color (same vs. different) across consecutive trials, separately for the three inter-trial target location transitions (target at previous target location, TT vs. at previously empty neutral location, TN vs. at previous distractor location, TD). A repeated-measures ANOVA with color (repetition/switch), response (repetition/switch), and target position (TT, TN, TD) as factors revealed all main effects to be significant (response: F(1, 13) = 7.75, p = .015, = 0.37, BFincl > 1000; color: F(1, 13) = 160.9, p < .001, = 0.93, BFincl > 1000; position; F(1.2, 15.6) = 72.82, p < .001, = 0.85, BFincl > 1000, Huynh-Feldt corrected degrees of freedom). RTs were faster when the target-defining color repeated vs. changed (48-ms difference), and when the response-critical notch position repeated vs. changed (16-ms difference). And RTs were significantly faster when the target appeared at the same position compared to either a previous distractor location (TT vs. TD: 46-ms difference, Bonferroni-corrected t(13) = 9.28, pbonf < .001, BF10 > 1000) or a previously empty (neutral) location (TT vs. TN: 30-ms difference, t(13) = 7.15, pbonf < .001, BF10 > 1000); there was also a significant cost when the target appeared at a previous distractor position vs. a previously empty (neutral) position (TD vs. TN: –16-ms difference, t(13) = –9.48, pbonf < .001, BF10 > 1000). In addition, the interactions RCF × position (F(1.4, 19) = 23.3, p < .001, = 0.64, BFincl > 1000 and RCF × color (F(1,13) = 6.19, p = .027, = 0.32, BFincl = 0.56) were significant, although the RCF × color interaction was not supported by the Bayesian analysis. Repeated-measures ANOVAs conducted separately for each positional inter-trial transition condition with response (repetition/switch) and color (repetition/switch) as factors revealed that the effects involving response (position of the target notch) were significant only in the repeated target position (TT) condition (main effect of response: F(1,13) = 32.6, p < .001, = 0.72, BFincl > 1000, response × color interaction: F(1,13) = 6.0, p = .029, = 0.32, BFincl = 0.89), but not in the TN or TD conditions (all p > .3). Download: PPT PowerPoint slide PNG larger image TIFF original image Fig 3. Response times. Mean response times for repeated/non-repeated color, target notch ‘orientation’ (the response- critical feature), and for the different positional transitions: target at previous target position (TT), target at previously neutral (i.e., empty) position (TN), and target at previous distractor position (TD). Error bars show the 95% confidence intervals. RCF: response-critical feature. https://doi.org/10.1371/journal.pcbi.1009332.g003 Model comparison results In our modeling framework (see Fig 4), we treat each trial of the experiment as a perceptual decision, which is modeled as an evidence accumulation process, and we allow the parameters of that evidence accumulation process (i.e., the starting point, rate of evidence accumulation, and non-decision time) to change from trial to trial based on recent stimulus history. Download: PPT PowerPoint slide PNG larger image TIFF original image Fig 4. Illustration of the hierarchical modeling framework. Each hierarchical model consists of an evidence accumulation model (either the drift-diffusion model or the LATER model) and either no updating or one updating rule for each parameter of the evidence accumulation model (starting point S0, evidence accumulation rate r, and non-decision time τ). Each updating rule belongs to one of the four categories shown in the middle layer of the figure, and is applied to one of the three inter-trial updating variables shown in the blue box on the right side of the figure: the response-critical feature (RCF), the target-defining color, or the target position. For some of the updating rules based on the RCF or color, there are three different versions of the rule, differing in their degree of position-specificity (top level of the hierarchy). These rules could be fully position-independent (PI), fully position-dependent (PD), with a gradient-like dependence on the change of position (PG), or starting out fully position-dependent but then spreading (PS). See the main text for detailed descriptions. https://doi.org/10.1371/journal.pcbi.1009332.g004 In particular, we consider three aspects of stimulus history: the cross-trial history of the response-critical feature (RCF), of the target color, and of the target position. For each of these, we consider different updating rules, each implementing a possible way in which one model parameter could change based on this aspect of stimulus history (see detailed mathematical description in the section of “Models and updating rules”). The aspect of stimulus history that an updating rule is based on is referred to as the updating variable (UV) of that rule. Each of the updating rules belongs to one of four categories: Bayesian rules, step rules, weighted rules, and binary rules. Binary rules have two different values of the updated parameter, one on trials where the UV was repeated from the immediately preceding trial (n-1) and the other on trials where it changed. Binary rules were included for comparison, serving as a ‘baseline’ to assess how much better inter-trial effects could be explained when taking into account trial history further back than n-1; step rules, weighted rules, and Bayesian rules represent three different ways of doing this. Step rules assume that repetition effects were partially carried over to future trials. For example, if the evidence accumulation rate was faster because the target color had been repeated between trial n-1 and trial n, some of this repetition benefit would be carried over to trial n+1 (and a smaller proportion carried over to trial n+2 and so on). Weighted rules instead assume that each state of the UV had an associated weight which determined the value of the updated parameter on trials where that state of the UV occurred. After each trial, some of the weight was shifted to the state of the UV which occurred on that trial, in such a way that the total weight remained constant, as if a limited resource was being reallocated. There was also memory decay of previous weight reallocations. The Bayesian rules were applied specifically to the starting point parameter, and assumed that the relative frequencies of the different states of the UV were learned through Bayesian updating, with memory decay as in the dynamic belief model of Yu and Cohen [35]. These frequencies were assumed to define a prior for the evidence accumulation process on each trial, implemented by setting the starting point to the logarithm of the prior odds of the state of the UV on that trial. For some of the updating rules based on the RCF and the target color, we also compared four different versions of the rule, differing in their degree of position specificity as well as in whether memories learned at a particular position would spread over time. These rules could be fully position independent (PI), fully position-dependent (PD), with a gradient-like dependence on the change of position (PG), or initially fully position-dependent but with spreading over time (PS). The PI rule assumes that the influence from RCF or color on previous trials does not depend at all on whether the target position was the same or had changed, while the PD rule assumes no influence at all from a previous trial unless the same target position is repeated. By contrast, the rule of PG suggests a stronger influence from a previous trial the closer the target is to its previous position. Finally, the PS rule assumes that learning is initially fully position specific but later spreads (e.g. because the exact position in which a particular target appeared is gradually forgotten). For some of the position-based updating rules, there are three different versions, with and without inhibition of previous distractor locations which could be either fully matched with target location facilitation so that weight is only transferred from distractor to target locations, or involve separate processes of transferring weight to the target and away from distractors (including to and from empty locations). Further details about the updating rules and the model fitting are provided in the Methods and models section below. We compared different updating rules based on how well the model, when using this updating rule, predicts RTs on all trials and determined the best of these updating rules in terms of the Akaike Information Criterion (AIC; see Methods and models section for more details). The AIC is a measure of the quality of a model, which takes into account goodness of fit (as measured by the likelihood) and also penalizes models with more free parameters. Lower AIC values indicate better model performance. In total, we compared 12 different updating rules for updating based on the response-critical feature, eight for the target color and ten for the target position (taking into account also the two different evidence accumulation models DDM and LATER, there were thus a total of 2*12*8*10 = 1920 possible models). Figs 5, 6 and 7 show the mean relative Akaike Information Criteria (AIC) for each of the response feature-, target color-, and target position-related updating rules. For each of these three updating variables, the AIC for each updating rule was evaluated for a model which used the best updating rule, in terms of having the lowest associated AIC, for each of the other two updating variables (see Methods and models section for further details). For each individual participant and session, we subtracted the AIC of the overall winning model (based on all participants and sessions) from the AIC of every other model for that participant and session, and finally we averaged this relative AIC across all participants and sessions. This resulted in a relative AIC of zero for the winning model, while the relative AIC of other models indicate how much worse they are compared to the winning model. Download: PPT PowerPoint slide PNG larger image TIFF original image Fig 5. Model comparison for color-based updating. Mean relative AIC for all different color-based updating rules when using the best updating rules for response-critical feature (RCF) and position. The dashed vertical line separates rules that update based on color alone (left) and rules that also take the target position into account (right). The different background colors mark rules that update either the rate (orange) or the non-decision time (green). Error bars represent the standard error of the mean over participants and sessions. PI: position-independent; PD: full position-dependent; PG: a gradient-like dependence on the change of position; NDT: Non-decision time. https://doi.org/10.1371/journal.pcbi.1009332.g005 Download: PPT PowerPoint slide PNG larger image TIFF original image Fig 6. Model comparison for position-based updating. Mean relative AIC for all different position-based updating rules when using the best updating rules for response-critical feature (RCF) and color. The different background colors mark rules that update either the rate (orange) or the non-decision time (green). Error bars represent the standard error of the mean over participants and sessions. DI: distractor inhibition; NDT: Non-decision time. https://doi.org/10.1371/journal.pcbi.1009332.g006 Download: PPT PowerPoint slide PNG larger image TIFF original image Fig 7. Model comparison for response feature-based updating. Mean relative AIC for all different RCF-based updating rules when using the best updating rules for color and position. The dashed vertical line separates rules that update based on the RCF alone (left) and rules that also take the target position into account (right). The different background colors mark rules that update either the starting point (blue), the rate (orange) or the non-decision time (green). Error bars represent the standard error of the mean over participants and sessions. PI: position-independent; PD: fully position-dependent; PG: a gradient-like dependence on the change of position; NDT: Non-decision time; S0: Starting point. https://doi.org/10.1371/journal.pcbi.1009332.g007 Target color-based updating models. To find the best rule for target color-based updating, we compared the different target color-based updating rules in terms of the mean relative AIC (see Fig 5). The best rule was the position-independent weighted-rate rule (see updating rule 7 in the “Models and updating rules” subsection of the Methods and models section for details). This updating rule assumes that the evidence accumulation rate depends on how much of a limited “weight” resource, which is distributed across both of the possible target colors, is allocated to the target color. This weight is updated after each trial by shifting some weight to the target color on that trial, with partial “forgetting” of old updates (see Fig 8 for an example, Fig 9 for the predicted temporal profile with this updating rule, and Methods and models section for more details). Download: PPT PowerPoint slide PNG larger image TIFF original image Fig 8. Example of color-based updating. Example of the weight changes to the different colors predicted by the weighted-rate updating rule for the first eight trials of a typical participant (A), and the associated changes in the evidence accumulation rate on the same eight trials (B). The letters “T” and “D” denote the target color and, respectively, the distractor color on each trial. The dashed line marks the baseline evidence accumulation rate. The evidence accumulation rate on each trial was the baseline rate scaled by the weight associated with the target position on that trial. https://doi.org/10.1371/journal.pcbi.1009332.g008 Download: PPT PowerPoint slide PNG larger image TIFF original image Fig 9. Temporal profile of the color-based inter-trial effects. Mean normalized RT for repeated vs. switched target color on (current) trial n compared to (preceding) trial n-1 (lag 1), n-2 (lag 2) up to n-8 (lag 8). Normalized RTs were first averaged across the two sessions for each participant; the resulting (individual mean normalized) RTs were then used to compute the overall means and confidence intervals, across participants. Filled circles depict the behavioral data, lines the model predictions. Error bars represent 95% confidence intervals. https://doi.org/10.1371/journal.pcbi.1009332.g009 Target position-based updating models. Next, we compared the mean relative AIC of position-based updating rules (see Fig 6). The best rule for position-based updating was the weighted-rate with distractor inhibition rule (see updating rule 8 in the “Models and updating rules” subsection of the Methods and models section for details). This updating rule assumes that the evidence accumulation rate depends on how much of a limited weight resource, which is distributed across all the possible target positions, is allocated to the target position. This weight is updated after each trial by shifting weight to the target position on that trial from all other positions, both distractor positions and empty positions, and shifting weight away from the distractor positions on that trial to all other positions, both target positions and empty positions, with partial forgetting of old updates (see Fig 10 for an example and Fig 11 for the predicted temporal profile with this updating rule, and Methods and models section for more details). Download: PPT PowerPoint slide PNG larger image TIFF original image Fig 10. Example of position-based updating. Example of the weight changes to the different positions predicted by the “weighted rate with distractor inhibition” updating rule for the first eight trials of a typical participant (A), and the associated changes in the evidence accumulation rate on the same eight trials (B). The letters “T” and “D” denote the target position and, respectively, the positions of the three distractors on each trial. The dashed line marks the baseline evidence accumulation rate. The evidence accumulation rate on each trial was the baseline rate scaled by the weight associated with the target position on that trial. https://doi.org/10.1371/journal.pcbi.1009332.g010 Download: PPT PowerPoint slide PNG larger image TIFF original image Fig 11. Temporal profile of the position based inter-trial effects. Mean normalized RT for different positional inter-trial conditions, target in previous target condition (TT), target in previous distractor position (TD), target in previously neutral (unoccupied) position (TN), compared to trial n-1 (lag 1), n-2 (lag 2) up to n-8 (lag 8). Normalized RTs were first averaged across the two sessions for each participant; the resulting (individual mean normalized) RTs were then used to compute the overall means and confidence intervals, across participants. Filled circles show the behavioral data, while lines show model predictions. Error bars represent 95% confidence intervals. https://doi.org/10.1371/journal.pcbi.1009332.g011 RCF-based updating models. Finally, we compared the mean relative AIC of response-based updating rules (see Fig 7). The best rule was “Position Gradient (PG) Bayesian S0” (see Fig 12 for an example and Fig 13 for the predicted temporal profile and updating rule 4 in the “Models and updating rules” subsection of the Methods and models section for details). This updating rule learns a different starting point for each possible target position, updating the starting point primarily for the actual position where the target occurred, though with some updating carried over to other positions, with the magnitude of the update decreasing with distance (we refer to this a gradient-like dependence on the change of position). Interestingly, both the positionally non-specific version of the Bayesian starting point updating rule and the completely position-specific version (which updates only a single position with no effect on other positions) performed considerably worse. Download: PPT PowerPoint slide PNG larger image TIFF original image Fig 12. Example of response-based updating. Example of the starting point changes associated with the different positions predicted by the gradient-like dependence on the change of position (i.e., “PG Bayesian S0” updating rule) for the first eight trials of a typical participant (A), and the starting point for the target position on the same eight trials (B). The position of the triangles represent the target position on each trial, and the shape of the triangle indicates the response-critical feature (RCF, i.e., whether the notch was on the top or bottom of the target item). https://doi.org/10.1371/journal.pcbi.1009332.g012 Download: PPT PowerPoint slide PNG larger image TIFF original image Fig 13. Temporal profile of the response feature-based inter-trial effects. Mean normalized RT for repetition vs. switch of the response defining target feature (notch on top or bottom of the diamond shape), compared to trial n-1 (lag 1), n-2 (lag 2) up to n-8 (lag 8) for the different positional inter-trial conditions, target in previous target condition (TT), target in previous distractor position (TD), target in previously neutral (unoccupied) position (TN). Normalized RTs were first averaged across the two sessions for each participant; the resulting (individual mean normalized) RTs were then used to compute the overall means and confidence intervals, across participants. Filled circles show the behavioral data, while lines show model predictions. Error bars represent 95% confidence intervals. RCF: response-critical feature. https://doi.org/10.1371/journal.pcbi.1009332.g013 Cross-validation. In order to examine to what extent the better fit of the best updating rules compared to the ‘No-update’ rule was a result of overfitting, we performed cross-validation on the best model and the no-update model. Specifically, since each session consisted of eight blocks of trials, we performed an eight-fold cross-validation, that is: for each block, we evaluated the model’s prediction in that (‘test’) block after training the model on the remaining seven (‘training’) blocks (the stimulus sequence in the test block was still used for updating, so that the starting point and weight would start with the correct values at the beginning of the next block, but participants’ performance in that block was not used for optimizing the model parameters). Table 1 presents the average log-likelihood (logarithm of the likelihood) for the best model as well as the model with no updating, each evaluated on the training set (averaging the log-likelihood in each block across the seven folds in which that block was included in the training set) as well as on the test set. The average log-likelihood was somewhat worse when evaluated on the test set, indicative of some degree of overfitting, and this difference was larger for the best model compared to the no-update model, suggesting there was some overfitting of the updating rules, in addition to the evidence accumulation model. Importantly, however, even when evaluated on the test set, the log-likelihood was substantially better for the best model compared to the no-update model, indicating that the updating rules did capture patterns in the data that generalize across blocks. This conclusion is further supported by the cross-validated out-of-fold predictions for the temporal profiles of the inter-trial effects presented in S6 Appendix, which fit the pattern in the data nearly as well as the corresponding predictions without cross-validation (depicted in Figs 9, 11 and 13 below). Download: PPT PowerPoint slide PNG larger image TIFF original image Table 1. Average log-likelihood, across participants and sessions, for the best model and the no-update model, evaluated on the training set as well as on the test set, and the difference between the test and training set log-likelihood (right column). The bottom column shows the difference in log-likelihood between the best model and the no-update model. https://doi.org/10.1371/journal.pcbi.1009332.t001 Target color-based updating models. To find the best rule for target color-based updating, we compared the different target color-based updating rules in terms of the mean relative AIC (see Fig 5). The best rule was the position-independent weighted-rate rule (see updating rule 7 in the “Models and updating rules” subsection of the Methods and models section for details). This updating rule assumes that the evidence accumulation rate depends on how much of a limited “weight” resource, which is distributed across both of the possible target colors, is allocated to the target color. This weight is updated after each trial by shifting some weight to the target color on that trial, with partial “forgetting” of old updates (see Fig 8 for an example, Fig 9 for the predicted temporal profile with this updating rule, and Methods and models section for more details). Download: PPT PowerPoint slide PNG larger image TIFF original image Fig 8. Example of color-based updating. Example of the weight changes to the different colors predicted by the weighted-rate updating rule for the first eight trials of a typical participant (A), and the associated changes in the evidence accumulation rate on the same eight trials (B). The letters “T” and “D” denote the target color and, respectively, the distractor color on each trial. The dashed line marks the baseline evidence accumulation rate. The evidence accumulation rate on each trial was the baseline rate scaled by the weight associated with the target position on that trial. https://doi.org/10.1371/journal.pcbi.1009332.g008 Download: PPT PowerPoint slide PNG larger image TIFF original image Fig 9. Temporal profile of the color-based inter-trial effects. Mean normalized RT for repeated vs. switched target color on (current) trial n compared to (preceding) trial n-1 (lag 1), n-2 (lag 2) up to n-8 (lag 8). Normalized RTs were first averaged across the two sessions for each participant; the resulting (individual mean normalized) RTs were then used to compute the overall means and confidence intervals, across participants. Filled circles depict the behavioral data, lines the model predictions. Error bars represent 95% confidence intervals. https://doi.org/10.1371/journal.pcbi.1009332.g009 Target position-based updating models. Next, we compared the mean relative AIC of position-based updating rules (see Fig 6). The best rule for position-based updating was the weighted-rate with distractor inhibition rule (see updating rule 8 in the “Models and updating rules” subsection of the Methods and models section for details). This updating rule assumes that the evidence accumulation rate depends on how much of a limited weight resource, which is distributed across all the possible target positions, is allocated to the target position. This weight is updated after each trial by shifting weight to the target position on that trial from all other positions, both distractor positions and empty positions, and shifting weight away from the distractor positions on that trial to all other positions, both target positions and empty positions, with partial forgetting of old updates (see Fig 10 for an example and Fig 11 for the predicted temporal profile with this updating rule, and Methods and models section for more details). Download: PPT PowerPoint slide PNG larger image TIFF original image Fig 10. Example of position-based updating. Example of the weight changes to the different positions predicted by the “weighted rate with distractor inhibition” updating rule for the first eight trials of a typical participant (A), and the associated changes in the evidence accumulation rate on the same eight trials (B). The letters “T” and “D” denote the target position and, respectively, the positions of the three distractors on each trial. The dashed line marks the baseline evidence accumulation rate. The evidence accumulation rate on each trial was the baseline rate scaled by the weight associated with the target position on that trial. https://doi.org/10.1371/journal.pcbi.1009332.g010 Download: PPT PowerPoint slide PNG larger image TIFF original image Fig 11. Temporal profile of the position based inter-trial effects. Mean normalized RT for different positional inter-trial conditions, target in previous target condition (TT), target in previous distractor position (TD), target in previously neutral (unoccupied) position (TN), compared to trial n-1 (lag 1), n-2 (lag 2) up to n-8 (lag 8). Normalized RTs were first averaged across the two sessions for each participant; the resulting (individual mean normalized) RTs were then used to compute the overall means and confidence intervals, across participants. Filled circles show the behavioral data, while lines show model predictions. Error bars represent 95% confidence intervals. https://doi.org/10.1371/journal.pcbi.1009332.g011 RCF-based updating models. Finally, we compared the mean relative AIC of response-based updating rules (see Fig 7). The best rule was “Position Gradient (PG) Bayesian S0” (see Fig 12 for an example and Fig 13 for the predicted temporal profile and updating rule 4 in the “Models and updating rules” subsection of the Methods and models section for details). This updating rule learns a different starting point for each possible target position, updating the starting point primarily for the actual position where the target occurred, though with some updating carried over to other positions, with the magnitude of the update decreasing with distance (we refer to this a gradient-like dependence on the change of position). Interestingly, both the positionally non-specific version of the Bayesian starting point updating rule and the completely position-specific version (which updates only a single position with no effect on other positions) performed considerably worse. Download: PPT PowerPoint slide PNG larger image TIFF original image Fig 12. Example of response-based updating. Example of the starting point changes associated with the different positions predicted by the gradient-like dependence on the change of position (i.e., “PG Bayesian S0” updating rule) for the first eight trials of a typical participant (A), and the starting point for the target position on the same eight trials (B). The position of the triangles represent the target position on each trial, and the shape of the triangle indicates the response-critical feature (RCF, i.e., whether the notch was on the top or bottom of the target item). https://doi.org/10.1371/journal.pcbi.1009332.g012 Download: PPT PowerPoint slide PNG larger image TIFF original image Fig 13. Temporal profile of the response feature-based inter-trial effects. Mean normalized RT for repetition vs. switch of the response defining target feature (notch on top or bottom of the diamond shape), compared to trial n-1 (lag 1), n-2 (lag 2) up to n-8 (lag 8) for the different positional inter-trial conditions, target in previous target condition (TT), target in previous distractor position (TD), target in previously neutral (unoccupied) position (TN). Normalized RTs were first averaged across the two sessions for each participant; the resulting (individual mean normalized) RTs were then used to compute the overall means and confidence intervals, across participants. Filled circles show the behavioral data, while lines show model predictions. Error bars represent 95% confidence intervals. RCF: response-critical feature. https://doi.org/10.1371/journal.pcbi.1009332.g013 Cross-validation. In order to examine to what extent the better fit of the best updating rules compared to the ‘No-update’ rule was a result of overfitting, we performed cross-validation on the best model and the no-update model. Specifically, since each session consisted of eight blocks of trials, we performed an eight-fold cross-validation, that is: for each block, we evaluated the model’s prediction in that (‘test’) block after training the model on the remaining seven (‘training’) blocks (the stimulus sequence in the test block was still used for updating, so that the starting point and weight would start with the correct values at the beginning of the next block, but participants’ performance in that block was not used for optimizing the model parameters). Table 1 presents the average log-likelihood (logarithm of the likelihood) for the best model as well as the model with no updating, each evaluated on the training set (averaging the log-likelihood in each block across the seven folds in which that block was included in the training set) as well as on the test set. The average log-likelihood was somewhat worse when evaluated on the test set, indicative of some degree of overfitting, and this difference was larger for the best model compared to the no-update model, suggesting there was some overfitting of the updating rules, in addition to the evidence accumulation model. Importantly, however, even when evaluated on the test set, the log-likelihood was substantially better for the best model compared to the no-update model, indicating that the updating rules did capture patterns in the data that generalize across blocks. This conclusion is further supported by the cross-validated out-of-fold predictions for the temporal profiles of the inter-trial effects presented in S6 Appendix, which fit the pattern in the data nearly as well as the corresponding predictions without cross-validation (depicted in Figs 9, 11 and 13 below). Download: PPT PowerPoint slide PNG larger image TIFF original image Table 1. Average log-likelihood, across participants and sessions, for the best model and the no-update model, evaluated on the training set as well as on the test set, and the difference between the test and training set log-likelihood (right column). The bottom column shows the difference in log-likelihood between the best model and the no-update model. https://doi.org/10.1371/journal.pcbi.1009332.t001 Temporal profiles of the inter-trial effects and model predictions Having seen which updating rule best accounts for each type of inter-trial effect, in this section we will explore what prediction those updating rules make for the temporal profile of each inter-trial effect and how well this matches the temporal profiles in the experimental data. In the next three subsections, we explore how well the model framework predicts the three different types of inter-trial effect: (i) inter-trial effects related to repetition/switch of the target-defining color, more precisely: repetition of both the target and distractor colors versus swapping of the target and distractor colors; (ii) inter-trial effects related to the target position, for which there are three different transition conditions: target at previous target position, target at previously unoccupied (neutral) position, and target at previous distractor position; (iii) inter-trial effects related to the response-critical feature, the position (top or bottom) of the notch in the target item. All figures in the next three sections show normalized RTs, where the overall mean RT has been subtracted from the mean in each condition, for each participant and session (i.e., a positive normalized RT means that the response time is slower than overall average in a particular condition, while a negative normalized RT means it is faster than overall average). To make the model predictions directly comparable to the behavioral data, the mean RTs (from the behavioral data) were also subtracted from the model predictions. All model predictions are based on the best model, that is, using the factor level that resulted in the lowest AIC for each of the factors (see the model comparison section above); accordingly, the relevant figures below (Figs 9, 11 and 13) are all based on the exact same model (see S2 Appendix for the parameters of the model fits), applying all of the winning updating rules (even though each of these figures only illustrates the effects of one, or two in the case of Fig 13, of these rules). Color-based inter-trial effects. Fig 8 illustrates the predictions of the best updating rule for color-based inter-trial effects, that is, the “weighted-rate” rule, for the first eight trials of a typical participant. The two colors start out receiving equal weight (each with a weight of 1 so that the rate for each color is the baseline rate), and after each trial some weight is shifted to the target color on that trial, while additionally there is partial forgetting so that the weights are partially reset to their starting values (see modeling part of Methods and models section for details). The figure illustrates how repetition benefits arise in the model: On trials 2, 3, and 5, the same target color is repeated as on the previous trial; on these trials, the weight associated with the target color is larger than one because some weight has just been shifted to that color; on trials 4 and 6–8, by contrast, the target color has changed from the previous trial, resulting in weight of less than one associated with the target color because some weight was just shifted away from that color. The figure also illustrates influences from more than one trial back: the weight associated with the target color is larger on trial 3 compared to trial 2, because this is the second repetition in a row and the weight shifts accumulate; by contrast, the weight is considerably lower on trial 5 because this color repetition was preceded by a sequence of three trials with the other target color. The color-based inter-trial effects were predicted well by this updating rule (shown in Fig 9). Response times are faster (by 56 ms, on average) when the color from trial n-1 is repeated on trial n. The same pattern is found for the repetition versus switch of the color from more than one trial back, although the magnitude of the effect decreases with increasing lag. This pattern is captured well by the model, although the model somewhat underestimates the effect for longer lags (i.e., for influences from four or more trials back), while it, if anything, slightly overestimates the effect at the shortest lags (influences from one to three trials back). Position-based inter-trial effects. The best updating rule for target position-based inter-trial effects (illustrated in Fig 10), that is, the “weighted rate with distractor inhibition” rule, followed a similar logic to that for the color-based effects. All eight positions start with equal weight (each with a weight of 1 so that the rate for each position is the baseline rate), and after each trial some weight is shifted to the target position on that trial from all other positions (i.e., both distractor positions and unoccupied positions), and away from the distractor positions on that trial to all other positions (i.e., both the target position and unoccupied positions; see modeling part of Methods and models section for details). This results in increased weight associated with the previous target position and decreased weight associated with the previous distractor positions, while the weight associated with the neutral positions does not change much because they both lose weight to the target position and receive weight from the distractor positions. This explains why the model predicts facilitation of response times for “target at previous target position” transitions (TT) and a cost for “target to previous distractor position” transitions (TD) compared to when the target moves to a previously neutral (empty) position (TN). The figure also illustrates influences from more than one trial back: both trial 4 and trial 7 are TD transitions, but the weight associated with the target position is much lower on trial 7, because the distractors had appeared at that position an additional two times in the recent trial history. The model also predicts well for the position-based inter-trial effects (Fig 11). Compared to when the target on trial n appeared at a position which was neutral (unoccupied) on trial n-1 (TN), RTs were faster (by 36 ms, on average) when the target appeared at the previous target position (TT), and slower (by 17 ms) when the target appeared at a previous distractor position. The same pattern of positional inter-trial effects is seen for transitions from more than one trial back, though the size of the effect decreased with increasing lag. The model predicts this pattern, although it predicts a slightly larger target-target transition benefit relative to the distractor-target transition cost compared to the behavioral data. Response-based inter-trial effects. The best updating rule for the response-based inter-trial effects: “PG Bayesian S0” (illustrated in Fig 12) assumed that participants learned a separate response bias (starting point of the evidence accumulation model) for each position, based on the frequency with which each response has recently been associated with that position, but that updating of the bias in the current target position on any trial partially carried over to other positions in a gradient-like way. In the example shown in Fig 12, the starting point is biased towards the boundary associated with a “notch on top” response after the first trial, where the strongest bias is associated with the position where the target occurred but with some bias also for other positions. On the second trial, the target position was repeated but the notch position changed–so the bias learned from the first trial is nearly cancelled. However, because of the memory decay, the new notch position had a somewhat stronger influence than the old one; so, instead of complete cancellation, the result was a small bias in the other direction. The next five trials all had the notch on the top, but with the target occuring in different positions, resulting in a build-up of a bias towards that response associated with all positions. Because inter-trial effects for the response primarily occurred when the target position was repeated (see Fig 3), the RCF-based inter-trial effects are shown separately for the different positional inter-trial transition conditions (Fig 13). The model predicts well in the repeated target position condition (TT). However, for the other conditions, the model slightly overestimates the inter-trial effects from one trial back and somewhat underestimates those for longer lags. This happens because in the behavioral data, there are virtually no effects of repetition/switch of the RCF from a single trial back when the target position does not repeat, while there are some such effects from two and more trials back. Color-based inter-trial effects. Fig 8 illustrates the predictions of the best updating rule for color-based inter-trial effects, that is, the “weighted-rate” rule, for the first eight trials of a typical participant. The two colors start out receiving equal weight (each with a weight of 1 so that the rate for each color is the baseline rate), and after each trial some weight is shifted to the target color on that trial, while additionally there is partial forgetting so that the weights are partially reset to their starting values (see modeling part of Methods and models section for details). The figure illustrates how repetition benefits arise in the model: On trials 2, 3, and 5, the same target color is repeated as on the previous trial; on these trials, the weight associated with the target color is larger than one because some weight has just been shifted to that color; on trials 4 and 6–8, by contrast, the target color has changed from the previous trial, resulting in weight of less than one associated with the target color because some weight was just shifted away from that color. The figure also illustrates influences from more than one trial back: the weight associated with the target color is larger on trial 3 compared to trial 2, because this is the second repetition in a row and the weight shifts accumulate; by contrast, the weight is considerably lower on trial 5 because this color repetition was preceded by a sequence of three trials with the other target color. The color-based inter-trial effects were predicted well by this updating rule (shown in Fig 9). Response times are faster (by 56 ms, on average) when the color from trial n-1 is repeated on trial n. The same pattern is found for the repetition versus switch of the color from more than one trial back, although the magnitude of the effect decreases with increasing lag. This pattern is captured well by the model, although the model somewhat underestimates the effect for longer lags (i.e., for influences from four or more trials back), while it, if anything, slightly overestimates the effect at the shortest lags (influences from one to three trials back). Position-based inter-trial effects. The best updating rule for target position-based inter-trial effects (illustrated in Fig 10), that is, the “weighted rate with distractor inhibition” rule, followed a similar logic to that for the color-based effects. All eight positions start with equal weight (each with a weight of 1 so that the rate for each position is the baseline rate), and after each trial some weight is shifted to the target position on that trial from all other positions (i.e., both distractor positions and unoccupied positions), and away from the distractor positions on that trial to all other positions (i.e., both the target position and unoccupied positions; see modeling part of Methods and models section for details). This results in increased weight associated with the previous target position and decreased weight associated with the previous distractor positions, while the weight associated with the neutral positions does not change much because they both lose weight to the target position and receive weight from the distractor positions. This explains why the model predicts facilitation of response times for “target at previous target position” transitions (TT) and a cost for “target to previous distractor position” transitions (TD) compared to when the target moves to a previously neutral (empty) position (TN). The figure also illustrates influences from more than one trial back: both trial 4 and trial 7 are TD transitions, but the weight associated with the target position is much lower on trial 7, because the distractors had appeared at that position an additional two times in the recent trial history. The model also predicts well for the position-based inter-trial effects (Fig 11). Compared to when the target on trial n appeared at a position which was neutral (unoccupied) on trial n-1 (TN), RTs were faster (by 36 ms, on average) when the target appeared at the previous target position (TT), and slower (by 17 ms) when the target appeared at a previous distractor position. The same pattern of positional inter-trial effects is seen for transitions from more than one trial back, though the size of the effect decreased with increasing lag. The model predicts this pattern, although it predicts a slightly larger target-target transition benefit relative to the distractor-target transition cost compared to the behavioral data. Response-based inter-trial effects. The best updating rule for the response-based inter-trial effects: “PG Bayesian S0” (illustrated in Fig 12) assumed that participants learned a separate response bias (starting point of the evidence accumulation model) for each position, based on the frequency with which each response has recently been associated with that position, but that updating of the bias in the current target position on any trial partially carried over to other positions in a gradient-like way. In the example shown in Fig 12, the starting point is biased towards the boundary associated with a “notch on top” response after the first trial, where the strongest bias is associated with the position where the target occurred but with some bias also for other positions. On the second trial, the target position was repeated but the notch position changed–so the bias learned from the first trial is nearly cancelled. However, because of the memory decay, the new notch position had a somewhat stronger influence than the old one; so, instead of complete cancellation, the result was a small bias in the other direction. The next five trials all had the notch on the top, but with the target occuring in different positions, resulting in a build-up of a bias towards that response associated with all positions. Because inter-trial effects for the response primarily occurred when the target position was repeated (see Fig 3), the RCF-based inter-trial effects are shown separately for the different positional inter-trial transition conditions (Fig 13). The model predicts well in the repeated target position condition (TT). However, for the other conditions, the model slightly overestimates the inter-trial effects from one trial back and somewhat underestimates those for longer lags. This happens because in the behavioral data, there are virtually no effects of repetition/switch of the RCF from a single trial back when the target position does not repeat, while there are some such effects from two and more trials back. Discussion In the present study, we applied the evidence accumulation framework that we had previously used to model inter-trial effects in pop-out visual search tasks [12] to the data from a “priming of pop-out” study [19]. The primary aims were to examine, through modeling, the mechanisms underlying the temporal profile of n-back inter-trial effects on mean RTs and to investigate the degree of spatial specificity of the inter-trial effects for the response-critical target feature and the target-defining feature. Comparing 1920 different models (each possible combination of 2 evidence accumulation rules, 12 updating rules for the response, 8 updating rules for color and 10 for position) for each of 14 participants, we showed that the best models in general predicted the temporal profiles of the inter-trial effects well (see Figs 9, 11 and 13), with some interesting deviations that are discussed in more detail below. Feature-, position-, and response-based inter-trial dynamics in PoP Model comparisons suggest that the best model to predict the inter-trial effects related to the response-critical feature (RCF, i.e., notch position) is to update the starting point of the evidence accumulation process. This is consistent with what we found in our previous study [12], where a factorial model comparison revealed that in three pop-out visual search experiments, inter-trial effects for the RCF were best captured by updating of the starting point, both for a simple detection task (requiring a target-present/-absent response) and a target discrimination task (in which participants responded to the target-defining dimension, color/orientation). Interestingly, for the priming pop-out task, the best updating rule was the version of “position gradient” updating, which learns a different starting point for each target position, but updates multiple positions after each trial, with the magnitude of the update decreasing with distance from the target position. This performed better than either a completely position-independent rule which learns a single starting point for all positions, or a completely position-specific rule which learns a different starting point for each position and updates only the target position after each trial. Based on an analysis of inter-trial effects from a single trial back, the superiority of the “gradient” rule over the completely position-specific one may seem surprising, given that there was no evidence of any inter-trial effects related to the RCF except when the target position was repeated (see Fig 3). However, an analysis of the effect of the further-back history of the RCF and target position (see Fig 13) reveals the reason for the superiority of the “gradient” rule: repetition/switch of the RCF from two trials back or more (trial m