1. Introduction
Tests of fluid intelligence (
Gf) can be administered either untimed, or with a time constraint (usually at the test level, but sometimes as an item-level deadline: e.g.,
Kyllonen et al. 2018). Any investigator interested in measuring fluid intelligence has to decide between these two options. The choice is not an easy one, as it depends on how exactly measurement will be affected by time pressure.
Raven’s matrices, as the test most representative of fluid intelligence (
Carpenter et al. 1990), are a good illustration of the dilemma. On one hand, the test was explicitly designed to be completed untimed.
John C. Raven (
1938) noted that the progressive matrices “cannot be given satisfactorily with a time-limit”;
John Raven (
2008) remarked that “it would not make sense to set a time limit within which people have to show how high they can jump whilst also insisting that that they start by jumping over the lowest bar. Clearly, the most able would not be able to demonstrate their prowess […] it also follows that it makes no sense to time the test”.
On the other hand, a long testing time is an obstacle in many situations: a few participants have prolonged a testing session for over an hour trying to solve every single item in Raven’s Advanced Progressive Matrices (APM), which is psychologically interesting but logistically troublesome. This quickly led investigators to experiment with time limits (e.g.,
Bolton 1955). Short forms were developed (
Arthur and Day 1994;
Bilker et al. 2012;
Bors and Stokes 1998); various time limits were tested (
Hamel and Schmittmann 2006), and norms were ultimately made available for different time limits (
Raven et al. 1998). The end result is that as with most intelligence tests (
Wilhelm and Schulze 2002), in contemporary assessment, Raven’s matrices are often administered with a time constraint.
Is imposing time pressure a good or a bad thing? Time pressure has a limited detrimental effect on discriminating power (a reasonable time limit still allows most participants to finish most items, save for the final and most difficult items, which tend to have low success rates anyway; e.g.,
Bolton 1955), on reliability (e.g.,
Bolton 1955;
Poulton et al. 2022; see also
Hong and Cheng 2019), and on the dimensional structure (
Poulton et al. 2022) of Raven’s matrices. However, this limited impact on basic psychometric properties does not mean that versions with or without a time limit are equivalent (e.g.,
Davidson and Carroll 1945;
Rindler 1979). A more important question is whether time pressure impacts the validity of the task.
Time pressure can constitute a major threat to validity (
Lu and Sireci 2007); this point has been recognized for a long time (
Cronbach 1949). A speeded version of Raven’s matrices tends to correlate very well with the same task performed without a time limit (
Hamel and Schmittmann 2006), but this is not the only aspect of validity. Time pressure may affect the response processes which translate individual differences of reasoning ability into differences of performance (
Borsboom et al. 2004;
Borsboom and Mellenbergh 2007). In other words, if forcing participants to respond faster changes the way items are processed, in such a way that performance is less dependent on the reasoning processes the task is supposed to be measuring, then a time limit should not be used. A meta-analysis based on Raven’s matrices indicated that using a time limit substantially changes correlations between reasoning performance other constructs, suggesting that response processes are indeed affected by time pressure (
Tatel et al. 2020).
The literature has extensively covered various aspects of the effect of a time pressure on response processes and validity in intelligence tasks (e.g.,
Kyllonen and Zu 2016). Six main potential effects of a time pressure (and potential threats on task validity) can be listed: (1) preventing completion of certain items, (2) involving an additional contribution of mental speed, (3) constraining response times on items, (4) modifying aspects of cognitive processing of the items, (5) affecting psycho-affective variables such as test anxiety and motivation, and (6) differentially affecting individuals as a function of individual abilities (e.g., working memory). These potential effects of time pressure overlap to an extent (e.g., constraining response times may force qualitative changes in item processing).
2. Potential Effects of Time Pressure
2.1. Effect 1: Time Pressure Leads to Skipping Items
When performing an intelligence test under time pressure, some participants may lack enough time to finish the task. The task is then interrupted before completion, which means some items are never reached and never attempted by the participant, leading to a lower score. This means that a participant’s score no longer necessarily reflects their maximal level of reasoning performance (e.g.,
Goldhammer 2015), in the sense of the maximum number of problems they should have been able to solve given their level of intellectual ability (see also
Raven 2008).
This effect of time pressure on the omission of some problems has been the most discussed by classic psychometrics. It constitutes the basis of statistics that aim to summarize the effects of speededness based on the amount of items not reached by participants (e.g.,
Cronbach and Warrington 1951;
Gulliksen 1950b;
Stafford 1971). A similar rationale is implicit in factor analyses estimating a speededness factor based on the last, but not the first items (
Borter et al. 2020;
Estrada et al. 2017), in factor analyses assigning a loading on the speededness factor that increases with item serial position (e.g.,
Schweizer and Ren 2013), in attempts to estimate processing speed based on the number of omitted items (e.g.,
Schweizer et al. 2019a), and in the finding of poorer model fit for later items (
Oshima 1994).
One major challenge with the omission of certain items is that it could interact with test-taking strategies. Indeed, some participants may deliberately decide to spend enough time on early problems, with the risk of running out of time and having to skip later items, whereas others may prefer to proceed quickly throughout the whole test (
Goldhammer 2015;
Semmes et al. 2011). These test-taking strategies may possibly interact with individual differences, with more able participants being more skilled at managing their time and selectively speeding up or slowing down depending on item difficulty and remaining time (
van der Linden 2009). It is also noteworthy that some participants may choose to keep a margin of security, leading them not to use all the time they have available and finish a test or item before the deadline (see
Bolsinova and Tijmstra 2015). Conversely, there may be individual catch-up phenomena, so that participants speed on early items but selectively slow down later when they have time left on the counter.
2.2. Effect 2: Time Pressure Taps into a Speed Factor
Alternatively, some authors view processing speed as a fundamental component of intelligence (e.g.,
Vernon 1983): Jensen in particular speculated that processing speed could reflect basic differences at the neurological level, which could constitute a major underpinning of the general factor
g (
Jensen 1993,
1998). A related argument comes from the factor structure of intelligence: the Cattell–Horn–Carroll (CHC) theory of cognitive abilities explicitly includes speed factors as broad abilities under the general factor (
McGrew 2009;
Schneider and McGrew 2018; see also
McGrew 2023). This view makes mental speed an integral part of intelligence as a construct, and if mental speed is part of what we mean by “intelligence”, then forcing participants to work quickly should just tap into an additional dimension of intelligence, leaving task validity unaltered or even enhanced.
This argument has multiple problems, however. First, the observed correlation between mental speed and intelligence does not necessarily imply an important causal status for mental speed (e.g.,
Schubert et al. 2018), and it is doubtful whether mental speed actually has real-life implications that make it worth measuring (
Kyllonen and Zu 2016). Second, imposing a time limit and contaminating an intelligence test with speed-related variance can spuriously inflate correlations with other constructs also measured under time constraints (e.g.,
Ackerman et al. 2002;
Engle and Kane 2004;
Tatel et al. 2020). Third, although cognitive psychology often presents “mental speed” as a unitary ability, it is in fact a complex multidimensional construct (see
Danthiir et al. 2005;
Roberts and Stankov 1999; see also
Draheim et al. 2019, for a discussion of measurement issues). As a result, the CHC theory comprises multiple factors related to speed: processing speed in simple cognitive tasks (
Gs), reaction and decision speed for elementary single items (
Gt), speed in motor activities (
Gps), and rate and fluency for retrieval of information stored in long-term memory (
Gr). The relation between these factors (e.g., do they form a superordinate speed factor?) is currently unclear (
Schneider and McGrew 2018). Moreover, the speed at which a complex reasoning task can be performed does not map cleanly on any CHC factor and probably taps into a mix of
Gf and one or more of speed factors (including Gs, but also Gt in certain tasks, and possibly Gr which encompasses ideational fluency; see
Schneider and McGrew 2018). Fourth, speed is not solely a question of ability and also depends on motivation, personality, and an individual’s speed-accuracy tradeoff (
Shaw et al. 2020). Lastly, it is not even certain that the speed factor that appears under time constraints actually represents mental speed: in some cases, it may also reflect individual ability and individual strategies to deal with the time pressure (
Davison et al. 2012;
Semmes et al. 2011) or a different construct altogether such as a form of rule generation fluency (
Verguts et al. 1999). In short, imposing a time limit to a reasoning task and convoking a speed factor make the measure less tractable overall.
2.3. Effect 3: Time Pressure Constrains Response Times
Time pressure naturally encourages speeding in the task and therefore constrains the amount of time that can be spent on a given item. This may be viewed as a threat for validity or not, depending on whether a high speed of responding is taken as a reflection of high intelligence. As noted by
Schneider and McGrew (
2018), “the speed metaphor is often used in synonyms for
smart (e.g.,
quick-witted)”. In this view, it is inherently desirable to solve intellectual problems more quickly: if two participants have the same accuracy, it makes intuitive sense to believe that the faster one is more intelligent (
Thorndike et al. 1926). This approach considers speed as an integral aspect of performance in the task. One way to take this into account is to use composite scores that combine accuracy and speed (e.g.,
Bruyer and Brysbaert 2011;
Dennis and Evans 1996; another example is found in certain subtests of Wechsler scales, which give bonus points for quick answers) or to jointly model accuracy and response times (
Goldhammer and Kroehne 2014;
Klein Entink et al. 2009b).
With this perspective, the speed at which the response process is executed is an index of its effectiveness as much as the correctness of the response. Therefore, imposing a time limit and constraining time on task is not necessarily a problem (if the difficulties posed by problem complexity and limited time both challenge the same ability, then high-performing participants should be both faster and more accurate) and could even be viewed as an advantage (since a time limit constrains the response times of participants, this could make them more comparable in terms of accuracy: see
Goldhammer 2015; see also
Bolsinova and Tijmstra 2015).
However, this line of reasoning overlooks a critical aspect of solving complex intelligence tests: being fast is not necessarily a good thing. There are at least two ways to frame this idea. The first is to stress the fact that cognitive operations take time: limiting the amount of available time mechanically limits the number of operations that can be completed. Given that complex operations germane to fluid reasoning (such as rule induction) are constrained by simpler operations related to basic manipulation of information, time pressure is likely to affect complex operations to a greater extent (
Salthouse 1996). The other important point is that speed is not only an index of effective reasoning: a low speed also reflects carefulness (
Kyllonen and Zu 2016). In terms of cognitive processes, longer response times can largely reflect time spent for validation and evaluation of one’s response (
Goldhammer and Klein Entink 2011); one study showed that participants who care more about the results tend to respond more slowly (
Klein Entink et al. 2009a).
Empirical data have substantiated the idea that responding slowly can be positive. At the item level, an unpublished study of 159 participants with eye-tracking showed that longer fixations on a matrix problem were associated with better performance, which suggests that taking the time for reflection is beneficial (
de Winter et al. 2021). At the task level, RTs tend to be positively correlated with ability estimates, which means better participants tend to be slower (
DiTrapani et al. 2016;
Goldhammer and Klein Entink 2011;
Klein Entink et al. 2009b;
Partchev and De Boeck 2012). In the case when participants give fast responses, speed is especially negatively correlated with success rate (
Partchev and De Boeck 2012; note that this result was specific to Raven’s matrices and did not occur for a verbal analogies task).
Critically, the emphasis on slow responding appears to depend on ability and difficulty (
Goldhammer et al. 2014). Participants with a higher level of ability and/or motivation tend to modulate their RTs as a function of problem difficulty and spend much longer on difficult problems (
Perret and Dauvier 2018;
Gonthier and Roulin 2020; see also
Tancoš et al. 2023), suggesting that these require substantially more time to be solved correctly. In line with this view, the relation between RTs and accuracy is negative for easy problems but becomes less negative (
Dodonova and Dodonov 2013) or even positive for more difficult problems (
Becker et al. 2016;
Goldhammer et al. 2015). In terms of processing, it is likely that complex problems, which involve more logical rules and more components on which to apply these rules, require more time to elaborate a correct answer. In short, responding slowly can also be characteristic of high performance, especially for difficult problems and high-ability participants. It is also worth recalling that not all groups respond at the same speed: forcing fast responses may be more detrimental to participants with a slower response speed, such as young children (
Borter et al. 2020) and older adults (
Salthouse 1996).
2.4. Effect 4: Time Pressure Can Affect Cognitive Processing
Encouraging speeding when responding to a problem may conceivably affect cognitive processing, above and beyond limiting the amount of processing that can be performed. A few studies have even suggested that fast responses to an intelligence test involve a different ability or process than slow responses (
Partchev and De Boeck 2012;
DiTrapani et al. 2016), although no information was provided regarding the nature of this ability. There are multiple pathways by which cognitive processing could be affected.
At the item level, one possible way to conceptualize the possible effects of time pressure is to think of the response process in a reasoning task as a drift-diffusion model (e.g.,
Frischkorn and Schubert 2018;
Kang et al. 2022;
Lerche et al. 2020;
van der Maas et al. 2011). This class of models considers that when confronted with a problem, participants continuously accumulate evidence in a random walk process (modeled as a constant drift rate in the direction of the response, plus noise), until they reach a decision threshold. Encouraging participants to speed their responding due to a time limit could force them to lower their decision threshold, interfering with verification of their response as discussed in the previous section (
Goldhammer and Klein Entink 2011;
Klein Entink et al. 2009a;
Kyllonen and Zu 2016). This would translate as faster RTs, lower accuracy, and lower confidence in one’s response.
Apart from a change of decision threshold, time pressure could also force participants to accumulate information at a higher rate. Based on the decision-making literature, this could translate into several effects in terms of cognitive processing (
Johnson et al. 1993; see also
Ben Zur and Breznitz 1981;
Wright 1974), including acceleration (performing the same cognitive operations more quickly), filtration of information (considering less information before making a decision; see also
Salthouse 1996), or a change of strategy (tackling the task in a qualitatively different way). Acceleration or filtration would translate as faster responses in the task and lower accuracy; filtration in particular could also translate as lower accuracy conditional on RT, i.e., lower accuracy for the same RT, owing to the qualitatively different nature of information processing.
As for changes of strategy, there has been little study of the effects of time pressure on strategy use in intelligence tests, but such effects seem especially likely. Participants in complex learning tasks tend to switch to faster or more simple strategies under time pressure (see
Chuderski 2016); the same phenomenon is observed in mathematics tasks (
Caviola et al. 2017) and is assumed to occur in working memory tasks (
Friedman and Miyake 2004;
Lépine et al. 2005;
St Clair-Thompson 2007;
Thomassin et al. 2015). In the context of a matrix task, a change of strategy could mean turning away from the effective constructive matching strategy (
Chuderski 2016), which relies on the time-intensive process of reconstructing the correct answer by integrating all information in an item, to the less costly strategy of response elimination, which relies on testing each possible answer in turn to see if it seems to superficially fit the matrix (for a review, see
Laurence and Macedo 2022; see also
Bethell-Fox et al. 1984;
Snow 1980). There is also substantial evidence that participants often adopt a strategy of rapid guessing when under severe time constraints (
Attali 2005;
Jin et al. 2023;
Schnipke and Scrams 1997;
Schweizer et al. 2021), which would mean turning away from both constructive matching and response elimination. Critically, rapid guessing may not be constant across groups and across individuals (e.g.,
Must and Must 2013), providing another source of potential individual differences.
The effects of time pressure on cognitive processing of a given item may also go beyond what can be modeled at the item level: time pressure could also be expected to negatively affect learning, disrupting performance in a cumulative fashion over the course of the task. Learning is an important aspect of performance in Raven’s matrices: participants discover logical rules over simple items and then generalize them over more complex items presented later in the test (
Ren et al. 2014;
Verguts and De Boeck 2002), either explicitly or as a form of implicit or associative learning (
Ren et al. 2014). One study has suggested that time pressure is detrimental to learning in a matrix task (
Chuderski 2016), possibly because giving faster responses on early items means participants process logical rules more superficially, in a way that impedes transfer to more difficult items. This mechanism could contribute to selectively increasing the detrimental effect of time pressure on items presented towards the end of a test, although the particular design of this study (with participants completing two samples of items in the task in succession, without then with time pressure) makes it difficult to know if this effect would occur under more classic testing conditions.
2.5. Effect 5: Time Pressure Can Affect Anxiety and Motivation
Apart from direct effects due to the time restriction, it is also possible that the pressure itself has an effect on accuracy. Studies from the decision-making literature have suggested that participants perform worse under a time pressure, not only when there is an actual time restriction (
Cella et al. 2007) but also when there is a
perceived time pressure, even in the absence of any time manipulation (
DeDonno and Demaree 2008).
This phenomenon could be partly due to an effect of pressure on constructs related to intelligence: for instance, time pressure could decrease participant motivation to complete the task. One study showed that participants who had to complete a reasoning task under an explicit time pressure were less intrinsically motivated, as reflected in both lower ratings of interest and less time spent voluntarily engaging with the task materials after the end of the testing session (
Amabile et al. 1976). Under this view, time pressure could also conceivably change the relation between performance and motivation (see
Kuhn and Ranger 2015).
Perceived time pressure could also create stress or test anxiety in participants (e.g.,
Sussman and Sekuler 2022). This could interfere with performance in several ways, such as creating worrisome thoughts which use up resources in working memory (
Eysenck and Calvo 1992; for other examples, see
Ashcraft and Kirk 2001;
Moran 2016), although this mechanism is disputed (
Kellogg et al. 1999). This process has been mostly studied in the related contexts of academic achievement and math anxiety (
Caviola et al. 2017) and may also occur with intelligence tests. Time pressure could also conceivably interact with individual differences in anxiety: in the case of math reasoning, removing time pressure is sometimes observed to selectively increase performance for more anxious participants (
Plass and Hill 1986), although this is not always the case (
Kellogg et al. 1999; see also
Traub and Hambleton 1972).
2.6. Effect 6: Differential Effects of Time Pressure
Although time pressure does not seem to affect the relative position (rank-ordering) of participants to a large extent (
Preckel et al. 2011;
Vernon et al. 1985;
Wilhelm and Schulze 2002), time pressure could still be expected to interact with individual differences in ability in absolute terms so that the distance between high-ability and low-ability participants varies as a function of time pressure. A situation often observed in reasoning tasks is the choking under pressure effect, wherein imposing a pressure (such as instructions emphasizing the measurement of intelligence, the addition of social pressure, dual tasking, etc.) creates a larger decrement of performance for high-performing participants, especially those with high working memory capacity (WMC;
Gimmig et al. 2006; for examples with math tests, see
Beilock and Carr 2005;
Beilock and DeCaro 2007). Choking under pressure could also occur with time pressure, decreasing the distance between low- and high-ability participants.
The same effect could occur with WMC, instead of ability: time pressure has been observed to decrease the distance between low- and high WMC participants (
Colom et al. 2015), which could be problematic given that WMC is one of the major correlates of intelligence. On the other hand, the opposite effect has also been reported: it has been argued that speeded intelligence tests have higher correlations with WMC (
Chuderski 2013,
2015;
Tatel et al. 2020) because time pressure requires participants to integrate all information in working memory, leaving no time to decompose the problem. This would lead to time pressure
increasing the distance between low- and high-ability participants. This finding however was not replicated in other studies (
Colom et al. 2015; see also
Ren et al. 2018).
Apart from WMC, there is suggestive evidence that time pressure could increase the relation between performance in Raven’s matrices and spatial abilities (
Tatel et al. 2020). A differential effect of time pressure could also conceivably be found with other constructs, such as motivation: given that more motivated participants tend to spend longer on problems (e.g.,
Wise and Kong 2005), imposing a time pressure could selectively decrease the performance of participants with high motivation. Lastly, a differential effect could be found as a function of mental speed and more generally as a function of age: time pressure could disproportionately affect younger children with low mental speed (
Borter et al. 2020) and possibly older adults although this is not necessarily the case in practice (
Babcock 1994).
Given the fact that high-ability participants tend to modulate their RTs to spend selectively more time on more difficult items (
Gonthier and Roulin 2020;
Perret and Dauvier 2018;
Tancoš et al. 2023), all these possible differential effects might also be expected to interact with item difficulty: if time pressure affects high-ability participants to a larger extent, it may be even more true for the most difficult items. However, RT modulation in the face of difficulty is a relatively new topic in the literature, and this possibility has not been tested.