Hostname: page-component-745bb68f8f-f46jp Total loading time: 0 Render date: 2025-02-06T11:37:23.540Z Has data issue: false hasContentIssue false

Predicting real-time adaptive performance in a dynamic decision-making context

Published online by Cambridge University Press:  16 December 2014

Darren Good*
Affiliation:
Graziadio School of Business & Management, Pepperdine University, Los Angeles, CA, USA
Rights & Permissions [Opens in a new window]

Abstract

Individuals in organizations must frequently enact a series of ongoing decisions in real-time dynamic contexts. Despite the increasing need for individuals to manage dynamic decision-making demands, we still understand little about individual differences impacting performance in these environments. This paper proposes a new construct applicable to adaptation in such real-time dynamic environments. Cognitive agility is a formative construct measuring the individual capacity to exhibit cognitive flexibility, cognitive openness and focused attention. This study predicts that cognitive agility will impact adaptive performance in a real-time dynamic decision-making microworld computer game called the Networked Fire Chief; a simulation developed to study and train Australian fire fighters. Cognitive agility, operationalized through three distinct methods (performance measures, self-reports and external-rater reports), explained unique variance beyond measures of general intelligence on the total score of adaptive performance in the microworld.

Type
Research Article
Copyright
Copyright © Cambridge University Press and Australian and New Zealand Academy of Management 2014 

In today’s organizations managers operate in contexts of increasing change and complexity, leading to more dynamic decision making (DDM) (Kozlowski, Gully, Brown, Salas, Smith, & Nason, Reference Kozlowski, Gully, Brown, Salas, Smith and Nason2001). DDM is a type of decision making categorized by scenarios with real-time continual change, novelty, ambiguity and time constraints (Brehmer, Reference Brehmer1992). Scholars suggest responding to these demands through individual adaptability (Hesketh, Reference Hesketh1997; Pulakos, Arad, Donovan, & Plamondon, Reference Pulakos, Arad, Donovan and Plamondon2000; Baird & Griffin, Reference Boyatzis, Goleman and Rhee2006). Adaptability is generally referred to as an ability to change when necessary. However, scholars who explicitly contribute to the study of adaptability do not often address the complexity inherent in the cognitive aspects of adaptive performance, as it pertains to such real-time dynamic tasks.

Understanding an individual’s capacity to be adaptive at the cognitive level is a vital starting point to successfully navigate dynamic real-time environments (Glynn, Reference Glynn1996). In order to be adaptive in a real-time dynamic context, one must create a new understanding of information in the environment (Zaccaro, Reference Zaccaro2001), allow it to alter the course of thinking when necessary (Mitroff, Simons, & Franconeri, Reference Mitroff, Simons and Franconeri2002) and remain focused on relevant information (Lustig, May, & Hasher, Reference Lustig, May and Hasher2001). The purpose of this paper is to propose and test a new construct: cognitive agility, a cognitive capacity that supports real-time adaptive performance within the scope of a single DDM task.

A context that has proved useful to researchers studying cognition is the DDM microworld. Microworlds (real-time interactive computer-based simulations) mirror the complexity in real-life (Gonzalez, Vanyukov, & Martin, Reference Gonzalez, Vanyukov and Martin2005), presenting real-time decisions in a changing environment (Brehmer, Reference Brehmer1992). Microworlds offer an experimental alternative to the ‘paper and pencil’ tests that are often used in the assessment of ‘dynamic, interactive, and time oriented phenomena’ (DiFonzo, Hantula, & Bordia, Reference DiFonzo, Hantula and Bordia1998: 280). This study predicts that cognitive agility will lead to better performance in a DDM microworld. The variables that form cognitive agility will be measured using multiple methods (performance measures, self-reports and external-rater reports).

LITERATURE REVIEW

Theoretical foundations of cognitive agility

Environmental influences on cognition (like novelty and complexity) are not fixed but instead more perceptual, suggesting individual differences in how information is attended to, filtered, encoded and interpreted (Neisser, Reference Neisser1967). Given that managers face more change and complexity, an investigation into cognitive-related constructs that support adaptability in real-time dynamic environments is needed.

Adaptability is often used as an overarching term to describe a set of individual behaviors, leading to adaptation (Briscoe & Hall, Reference Briscoe and Hall1999). Therefore, it is important to unbundle the constituent concepts and seek greater clarity in both the concepts and measures as they relate to particular contexts. Aspects of adaptability have been interpreted as parts of various aspects of a person, including personality traits (Morrison, Reference Morrison1977), competencies (Boyatzis, Goleman, & Rhee, Reference Boyatzis, Goleman and Rhee1999), learning style (Mainemelis, Boyatzis, & Kolb, Reference Mainemelis, Boyatzis and Kolb2002) and cognitive style (Sternberg, Reference Sternberg1997). Others admit that, regardless of their definition or approach, adaptability is in part a cognitive capability (Mumford, Zaccaro, Harding, Jacobs, & Fleishman, Reference Mumford, Zaccaro, Harding, Jacobs and Fleishman2000).

While adaptability is a general ability to support change, cognitive agility is a specific cognitive capability applied to contexts that require a series of individual adaptations. The primary distinction comes from the possibility of categorizing adaptations along the dimensions of time and task. For instance, one can adapt in the moment, to real-time tasks (Lerch & Harter, Reference Lerch and Harter2001), or over a longer period of time, as in adjusting to a new job (Ashford & Taylor, Reference Baltes and Staudinger1990). In addition, one can adapt within a particular task that is changing (Cañas, Quesada, Antolí, & Fajardo, Reference Cañas, Quesada, Antolí and Fajardo2003) or across various tasks that make up a dynamic context, like adapting well to the introduction of a new technology (Edmondson, Bohmer, & Pisano, Reference Edmondson, Bohmer and Pisano2001).

Many studies predict abilities or characteristics of individual adaptability (Pulakos et al., Reference Pulakos, Arad, Donovan and Plamondon2000); most of which, at least implicitly, track adaptations across a longer time horizon and across multiple tasks. Fewer studies have focused on the cognitive aspects of adapting within a single real-time DDM task (exceptions include LePine, Colquitt, & Erez, Reference LePine, Colquitt and Erez2000; Cañas et al., Reference Cañas, Quesada, Antolí and Fajardo2003). Therefore, further inquiry into capabilities that predict real-time adaptive performance is a useful addition to the adaptability literature.

Real-time adaptive performance

Real-time adaptive performance within a task (i.e., how well an individual performs within a changing task, Kozlowski et al., Reference Kozlowski, Gully, Brown, Salas, Smith and Nason2001) likely requires a range of skills. A real-time DDM task context has change, novelty, ambiguity and complexity (Brehmer, Reference Brehmer1992). Individuals must enact a series of continuous decisions with various task-related tradeoffs (Tverskey, Sattath, & Slovic, Reference Tversky, Sattath and Slovic1988). To adapt one must be able to shift thinking when necessary (Mitroff, Simons, & Franconeri, Reference Mitroff, Simons and Franconeri2002) and doing so requires flexibility to override a dominant or automatic response in favor of a more appropriate one (Clark, Reference Clark1996).

Flexibility is preceded by the ability to notice stimuli of consequence. Yet, noticing too much information may cause unwanted distractions (Kuhl & Kazen-Saad, Reference Kuhl and Kazen-Saad1988) that can result in missing other relevant data (Shapiro & Raymond, Reference Shapiro and Raymond1994) and ultimately a loss in decision-making effectiveness (Anderson, Reference Baddeley and Hitch1983). Therefore, while it is important to notice relevant stimuli for task adaptability, adaptive decisions are also supported by the capacity to avoid less relevant data (Shanteau, Reference Shanteau1988).

Cognitive agility seeks to synthesize and evolve simultaneously the current conceptualizations of adaptability, adaptive performance and flexibility. Specifically, cognitive agility represents an individual’s cognitive capacity to flexibly operate with cognitive openness and focused attention. The following section provides theoretical reasoning for the choice of agility as a construct name as opposed to variations of adaptability or flexibility (i.e., other terms more commonly used to suggest a range of individual appropriate and variable behavior).

There are multiple ways in the literature to describe the capacity to make an appropriate change in response to the environment. Adaptability as described throughout this paper is perhaps the most common naming convention (Pulakos et al., Reference Pulakos, Arad, Donovan and Plamondon2000). Yet, adaptability is often operationalized as performance in a task that is complex, novel or ambiguous (LePine, Colquitt, & Erez, Reference LePine, Colquitt and Erez2000), regularly referred to as adaptive performance (Kozlowski et al., Reference Kozlowski, Gully, Brown, Salas, Smith and Nason2001). Therefore, describing adaptability as an ability leading to adaptive performance becomes tautological and potentially limiting toward extending knowledge of the construct. Some have labeled such adjustments as ‘cognitive adaptability,’ which means an ability to change decision frameworks or knowledge to meet the environmental needs (Haynie, Shepherd, Mosakowski, & Earley, Reference Haynie, Shepherd, Mosakowski and Earley2010). What this conceptualization may be missing with regard to the current study are the cognitive needs of the particular context. In real-time dynamic tasks, the speed of change is a necessary component, along with trying to identify what the individual is changing to and from in terms of orientation and decision frameworks. The context of a real-time DDM task is different than adapting on a longer time horizon, and therefore may represent a completely distinct class of decision-making performance that is separate from single decisions, requiring a different set of abilities (Sitkin & Pablo, Reference Sitkin and Pablo1992).

Cognitive flexibility is another construct used to signify appropriate adjustment to situational needs. It is saddled by a heterogeneity of definitions and operationalizations ranging from set-shifting to being creative, making it difficult to carefully conceptualize (Luchins & Luchins, Reference Luchins and Luchins1959; Isen, Daubman, & Nowicki, Reference Isen, Daubman and Nowicki1987; Spiro & Jehng, Reference Spiro and Jehng1990; Martin & Anderson, Reference Martin and Anderson1998; Cañas et al., Reference Cañas, Quesada, Antolí and Fajardo2003). In addition, flexibility is often used synonymously with adaptability (Pulakos et al., Reference Pulakos, Arad, Donovan and Plamondon2000), further confusing the two constructs. Statements from leading scholars on the subject such as, ‘cognitive flexibility is defined as a person’s willingness to be flexible and adapt to the situation’ are not at all uncommon (Martin & Anderson, Reference Martin and Anderson1998: 1). Again by using the terms ‘adapt’ and ‘flexible’ together, in place of one another, or to describe flexibility as being flexible, makes the understanding of the constructs less clear and contextually unbound.

Taken together, it is evident that cognitive flexibility is a construct with a conceptualization that has not been fully agreed upon. Yet, across the diverse definitions is a theme of intelligently adjusting to one’s environment (Berg & Sternberg, Reference Berg and Sternberg1985), through various forms of shifting, restructuring or expanding cognition. It is suggested here that the ability to control one’s thinking and change the decision strategy is just one aspect of cognition that allows for more adaptive performance in dynamic contexts (Cañas et al., Reference Cañas, Quesada, Antolí and Fajardo2003). Cognitive flexibility alone does not adequately describe what an individual is actually changing about his or her cognition. The particular change is likely relevant to specific environmental contexts. The elements encountered in a real-time dynamic context likely require a particular integration of flexibility with other necessary dimensions.

There is a need for identifying unique clusters of capacity that support adaptive performance in specific contexts. Therefore the word agility was chosen as it represents integration, coordination and a balance of multiple capabilities amidst changing conditions. It suggests an ability to do so quickly, which aligns with adapting in a real-time DDM. At the organizational level, agility describes the capacities of the firm to respond quickly to continual and ongoing environmental changes. Therefore cognitive agility is a construct at the individual-cognitive level that predicts adaptive performance within the specific context of a real-time dynamic task.

As a formative construct, cognitive agility includes the variables of cognitive openness, focused attention and cognitive flexibility (Figure 1). These three variables operate in unison within a task that demands dynamic real-time updates. In the particular microworld used (the Networked Fire Chief [NFC]) as well as others, intelligence has shown to predict success (Ackerman, Reference Anderson1992; Brehmer & Dörner, Reference Brehmer and Dörner1993; Gonzalez, Vanyukov, & Martin, Reference Gonzalez, Vanyukov and Martin2005). This study attempts to support these findings and extend them by demonstrating that cognitive agility, measured by multiple methods, impacts adaptive performance beyond intelligence.

Figure 1 Formative construct of cognitive agility

Intelligence and real-time adaptive performance

Intelligence is the ability to balance the demands of the situation to adapt with success (Sternberg, Reference Sternberg1999). General intelligence (g) refers to the individual capability to ‘broadly’ comprehend one’s environment and effectively plan a response (Gottfredson, Reference Gottfredson1997). Higher levels of g are associated with greater results in tasks that are novel and complex (Hunter & Hunter, Reference Hunter and Hunter1984; Ackerman, Reference Ackerman1988). Intelligence has been shown to predict performance outcomes in DDM scenarios (Gonzalez, Vanyukov, & Martin, Reference Gonzalez, Vanyukov and Martin2005) and g has been shown to predict adaptive performance in particular (LePine, Colquitt, & Erez, Reference LePine, Colquitt and Erez2000; Zaccaro, Reference Zaccaro2001).

The DDM scenario used in this study requires one to adapt in a novel and complex environment. As past DDM studies have shown that intelligence predicts success in the changing context(s) (Ackerman, Reference Anderson1992; Gonzalez, Vanyukov, & Martin, Reference Gonzalez, Vanyukov and Martin2005; Rigas, Carling, & Brehmer, Reference Rigas, Carling and Brehmer2002), it should predict higher performance in the particular DDM – a changing scenario with conflicting goals and numerous options to choose among (Brehmer, Reference Brehmer1995).

Focused attention and real-time adaptive performance

In dynamic situations one must be able to manage the effects of incoming information within the current course of cognitive action. With ongoing changes to the environment, focusing attention can fail owing to a situational occurrence that disrupts attention (Yantis, Reference Yantis1993; Theeuwes, Reference Theeuwes1994). Adaptive performance within a task requires that focus be paid to relevant data, which provide ongoing cues to adjust activity accordingly. If focused attention were low then an individual may find his or her attentional resources diluted in relation to an intended task. In essence, he or she may become overwhelmed by information, and less able to follow a coherent decision path toward adapting well with the changes (Anderson, Reference Baddeley and Hitch1983).

Cognitive openness and real-time adaptive performance

Cognitive openness as a concept is associated with several streams of existing literature, which are each, linked to adaptability. There is not an existing performance instrument or questionnaire to measure, explicitly, the cognitive aspects of being open. Therefore, the links between openness, creativity and mindfulness are established here to support the choice of terminology and measurement.

Openness is most often cited as ‘openness to experience,’ as used in Big Five personality trait measurements (Costa & McCrae, Reference Costa and McCrae1985). Still, other constructs such as creativity (Gough, Reference Gough1979), curiosity (Littman, Reference Littman2005) and mindfulness (Langer, Reference Langer1989), while they do not always employ the term openness per se, also highlight the importance of ‘being open’ or employing related qualities of cognition with respect to dynamic contexts. The term cognitive openness describes noticing novelty and creating new associations; a skill-related creativity. In fact, openness as used in the Big Five is often conceptually related to creativity to the point that scholars use the term creativity to refer to openness (Digman, Reference Digman1990; Matthews & Deary, Reference Matthews and Deary1998).

In a similar vein, Langer’s (Reference Langer1989) work on mindfulness, suggests a propensity to cognitively take in more aspects of a situation, allowing individuals to be more adaptive. One area within mindfulness, where this is most salient, is novelty seeking. Novelty seeking from a mindfulness conceptualization and measurement perspective includes aspects of openness, curiosity and creativity (Bodner, Reference Bodner2000). In a real-time dynamic context this combined resource supports one in seeking understanding (curiosity), embracing change (novelty/openness) and searching for ways to categorize information (creativity) that could lead to adaptive performance.

Cognitive flexibility and real-time adaptive performance

As previously mentioned, cognitive flexibility is necessary for adaptive performance within DDM contexts (Cañas et al., Reference Cañas, Quesada, Antolí and Fajardo2003). Compared with the earlier exploration of the more general definitions of cognitive flexibility, this study focuses on cognitive flexibility as an executive function that supports successful adaptation by regulating cognition and overriding routinized responses (Clark, Reference Clark1996).

Cognitive flexibility as conceptualized here, is in part a metacognitive regulative capability. Metacognition, commonly referred to as ‘thinking about thinking’ (Flavell, Reference Flavell1979), includes both the knowledge and regulation of cognitive activity (Moses & Baird, Reference Moses and Baird1999). The regulation component of metacognition is synonymous to cognitive flexibility, as it encompasses bottom-up processes like monitoring (e.g., error detection, source monitoring) and the top-down processes of cognitive control that include error correction, inhibitory control, planning and resource allocation (Reder & Schunn, Reference Reder and Schunn1996; Fernandez-Duque, Baird, & Posner, Reference Fernandez-Duque, Baird and Posner2000).

Cognitive agility and real-time adaptive performance

Adapting within a real-time dynamic task requires that one flexibly operate, being both open and focused. An individual must be able to notice novelty and have the capacity for the creation and inclusion of new information (Wallach & Kogan, Reference Wallach and Kogan1965; Kozlowski et al., Reference Kozlowski, Gully, Brown, Salas, Smith and Nason2001). Cognitive openness supports noticing relevant stimuli including aspects of the situation that may be easily overlooked (Langer, Reference Langer1989; Chan & Schmitt, Reference Chan and Schmitt2000). Yet, an over inclusiveness of information can hamper individuals who see many associations between ideas but have trouble focusing on one thing. Focused attention supports the decision maker in being able to resist the handling of less relevant information (Lustig, May, & Hasher, Reference Lustig, May and Hasher2001). Therefore, the individual must be able to do both. Cognitive flexibility, the ability to cognitively control and shift mental set (Rende, Reference Rende2000), then is also necessary in real-time adaptive performance. A combination of these three capacities, as a formative construct, should support adaptive performance through a real-time changing task context. In this study we measure cognitive agility using performance tests, self-reports and external-rater reports.

Hypothesis 1: Intelligence will explain significant variance in the adaptive performance score in the DDM scenario.

Hypothesis 2: The formative construct of cognitive agility measured by performance scores will explain significant variance in the adaptive performance score in the DDM scenario beyond that explained by intelligence.

Hypothesis 3: The formative construct of cognitive agility measured by self-reports will explain significant variance in the adaptive performance score in the DDM scenario beyond that explained by intelligence.

Hypothesis 4: The formative construct of cognitive agility measured by external-raters will explain significant variance in the adaptive performance score in the DDM scenario beyond that explained by intelligence.

METHODS

Participants

Undergraduate business students volunteered to be included on a contact list to take part in university-supported research studies. The overall response rate from the research contact list was 44% (420 email invitations were sent, 195 originally agreed to participate) with an effective response rate of 43% (181/420). The total useable sample consisted of 101 males and 80 females with a mean age of 21 years.

Procedure

The study had two parts. In Part I, participants were assigned to individual computer terminals in a lab setting where they completed three questionnaires, three performance tests and multiple trials of the microworld simulation (NFC). Participants were required to provide three email addresses of people who ‘know them best’ to take a survey on their behalf. Once Part I was completed, Part II began with an automated email sent to the persons the participant provided. The email connected to a survey that included the same three questionnaires the participants completed in Part I, but were reworded to collect information about how the selected ratee views the behavior of the participant (i.e., external-rater reports). Other than a change to the pronouns (e.g., from ‘I’ to ‘Him/Her’) the items remained exactly the same.

MEASURES

Control variables

The control variables were selected based on past empirical support that have used the same microworld simulation as a performance measure.

Verbal intelligence

The basic word vocabulary test is a 40-item test that measures the number of basic words that one knows (Dupuy, Reference Dupuy1974). The measure has high internal consistency close to 0.96 (Dupuy, Reference Dupuy1974). It correlates highly with other nationally standardized measures of verbal ability to include a 0.76 median correlation with the verbal sections of the Sequential Tests of Educational Progress and the School and College Abilities Tests (Dupuy, Reference Dupuy1974).

Visual–spatial intelligence

In the Card Rotations Test subjects look at a two-dimensional shape and decide whether eight figures to the right can be rotated to match the original image or are instead a mirror image of it. It is a 224-item test with a 6 min time limit. The test has shown reliability at 0.80 for males and 0.83 for females (Ekstrom, French, & Harman, Reference Ekstrom, French and Harman1976). The test correlates significantly with the Raven’s Progressive Matrices (r=0.40) (Pallier, Roberts, & Stankov, Reference Pallier, Roberts and Stankov2000), and the Short Form Test of Academic Aptitude (r=0.41) (Burns & Gallini, Reference Burns and Gallini1983). A total of correct answers is used as a measure of visual–spatial intelligence (Ekstrom, French, & Harman, Reference Ekstrom, French and Harman1976).

Independent variables

Cognitive openness self- and external-rater reports

Cognitive openness was measured using the 6-item Novelty Seeking subscale from the Langer Mindfulness Inventory (Bodner, Reference Bodner2000). An example item from this questionnaire is ‘I seek out new information even in a familiar situation.’ Those scoring highly on this scale are more likely to look for new information. This scale correlates with the Big Five factor model of ‘openness to experience’ (r=0.50), the Multiple Perspectives Inventory (r=0.64) and is negatively correlated with the need for cognitive closure (r=−0.20) (Bodner, Reference Bodner2000). The standardized factor loadings are 0.53 to 0.62 for the novelty seeking subscale (Bodner, Reference Bodner2000).

Cognitive openness performance score

The Alternate Uses Test (AUT) challenges the participant to list as many possible uses for a common item in a timed setting (i.e., a brick and paperclip used here; given 4 min/object; Guilford, Christensen, Merryfield, & Wilson, Reference Guilford, Christensen, Merrifield and Wilson1978). The number of items generated was used as the cognitive openness performance score. Reliability for the AUT has been demonstrated from 0.62 to 0.85 (AUT manual). The AUT correlates significantly with openness on the NEO (r=0.46; Chamorro-Premuzic, Reference Chamorro-Premuzic2006), the Barron Symbolic Equivalence Test (r=0.49; Barron, Reference Barron1988) and with greater sensitivity in a habituation process (r=0.36; Martindale, Anderson, Moore, & West, Reference Martindale, Anderson, Moore and West1996).

Focused attention self- and external-rater reports

Focused attention was assessed using the 9-item Focus of Attention subscale from The Attentional Control Scale (Derryberry & Rothbart, Reference Derryberry and Rothbart1988). High scores on the scale indicate a perceived capacity to limit the influences of irrelevant information from the environment. An example item from this scale is ‘When concentrating, I can focus my attention so that I become unaware of what’s going on in the room around me.’ The total scale correlates with the Inhibitory Control Scale (r=0.25) and with Trait Anxiety (r=−0.50) (Derryberry & Rothbart, Reference Derryberry and Rothbart1988). The measure is internally consistent (α=0.88).

Focused attention performance

The go/no go is a well-known paradigm for measuring focused attention (Zimmermann & Fimm, Reference Zimmermann and Fimm2000). Participants are challenged to respond (by pressing the space bar) as fast as possible when a ‘go’ stimuli appears and to withhold from responding (by not pressing the space bar) when ‘no go’ stimuli appear. Participants were asked to memorize two 3×3 textured squares (go stimuli). Then squares appear that are the same (go stimuli) and slightly different (no go stimuli). Reaction time (at the level of milliseconds) was used as the performance score. Reliability for go/no go has been demonstrated with split half and odd even coefficients at 0.998 (Zimmerman & Fimm, Reference Zimmermann and Fimm2000). Performance scores on the go/no go correlate with the Barrett Impulsiveness Scale (r=0.40, p<.01) and perseverative error on the Wisconsin Card Sorting Task (r=−0.46) (Keilp, Sackeim, & Mann, Reference Keilp, Sackeim and Mann2005).

Cognitive flexibility self- and external-rater reports

The 35-item Regulation of Cognition subscale, from the Metacognitive Awareness Inventory, was used as a proxy measure of cognitive flexibility (Schraw & Dennison, Reference Schraw and Dennison1994). The scale commonly assesses the ability to monitor and control strategies in goal-driven situations. An example item is, ‘I find myself-analyzing the usefulness of strategies while I make decisions.’ While explicit cognitive flexibility scales exist (e.g., Martin & Rubin, Reference Martin and Rubin1995), they do not conceptually express the control and appropriate shifting of cognition as cognitive flexibility is defined in this study (the Metacognitive Awareness Inventory is the most well-aligned survey found). The cognitive regulation subscale correlates with the Motivated Strategies for Learning Questionnaire on the Individual Learning Strategies Scale (0.72; Pintrich, Smith, Garcia, & McKeachie, Reference Pintrich, Smith, Garcia and McKeachie1991). The original Metacognitive Awareness Inventory has shown strong reliability at 0.90 (Schraw & Dennison, Reference Schraw and Dennison1994).

Cognitive flexibility performance score

The Stroop Task (Stroop, Reference Stroop1935) was used to measure cognitive flexibility as a performance score. The word-color Stroop task challenges the participant to respond to a font color in which an incongruent word of a color is presented (e.g., the word green written in red font). The participant is meant to select the word of the font color (red) instead of the word itself (green). As responding to the word rather than color is the automatic reaction (MacLeod, Reference MacLeod1991), it serves as a measure of the participant’s ability to override and flexibly respond. Performance was measured for 60 trials at the level of milliseconds. The Stroop Task correlates significantly with the Self-Monitoring Scale at r=0.43 and with demonstrated reliability at 0.86 (Koch, Reference Koch2003).

Dependent variable

Adaptive performance

The NFC is a real-time DDM microworld that was used to measure adaptive performance (Omodei & Wearing, Reference Omodei and Wearing1995). The NFC was originally developed with funding from the Australian Defence Force and the Australian Research Council to study real-time decision making as it occurs in real-world fire fighting. The NFC allows participants to ‘play’ the role of a ‘fire chief’ who extinguishes fires by delivering water carried by helicopters and fire trucks (see Figure 2). The microworld is constantly changing depending on various elements (e.g., fire spreading, flame intensity and wind direction/intensity).

Figure 2 Networked Fire Chief

While the NFC uses the context of fire fighting it is not meant to provide direct external validity. As with other DDM microworlds, it intends to simulate conditions of dynamism presented to individuals in external working conditions (Funke, Reference Funke1991; Brehmer & Dorner, Reference Brehmer and Dörner1993). The reason for using the microworld is to ‘map the functional relationships between the variables studied in the microworld, and not necessarily the surface similarities between the microworld and a particular field setting’ (DiFonzo, Hantula, & Bordia, Reference DiFonzo, Hantula and Bordia1998: 282). Microworlds are considered especially optimal for studies investigating highly complex phenomena in a dynamic context (Brehmer, Leplat, & Rasmussen, Reference Brehmer, Leplat and Rasmussen1991; Brehmer & Dorner, Reference Brehmer and Dörner1993). The NFC has demonstrated good reliability at 0.70 (Omodei et al., Reference Omodei, Wearing, McLennan, Hansen, Clancy, Elliott, Ley, Taranto and Thorsteinsson2001). Performance scores on the NFC have shown significant correlation with the core perceptual-cognitive characteristics of naturalistic decision making; namely, speed, accuracy, efficiency and planning (ranging from r=0.75, p<.01 for proactive planning to r=0.36, p<.05 for the percentage of time the closest appliance was allocated first; Elliot, Welsh, Nettelbeck, & Mills, Reference Elliot, Welsh, Nettelbeck and Mills2007). A performance score is produced by the program, which is calculated adding every safe cell (i.e., unburned). Demonstrating gains over losses after an individual has interacted with a changing environment is a way of calculating adaptive performance (Baltes & Staudinger, Reference Brehmer1996). A composite for the three trials (15 min in total) was used as a measure of adaptive performance.

RESULTS

Descriptive statistics

Table 1 shows descriptive statistics, bivariate correlations and reliabilities (in parentheses). All variables demonstrate reliability scores that are above 0.70. All variables with the exception of the external-rater reports for openness and focused attention correlate significantly or reach near significance with the dependent variable (DV). Contrary to expectation, the external-rater and self-report measures of cognitive flexibility were negatively correlated with the DV. This will be discussed in detail below and has implications for further analyses.

Table 1 Descriptive statistics

Note. N=181. Reliabilities are given in parentheses on the diagonal.

p<.1; *p<.05; **p<.01.

Factor analysis

Given the proposed formative nature of the cognitive agility construct, one way to further assess construct validity before regression analysis is through factor analysis (Rossiter, Reference Rossiter2002). Results of this analysis appear in Table 2. The factor analysis shows three factors, loading by method, with the exception of focused attention self-report, which loads weekly (0.331) on factor 3. Owing to the formative nature of the proposed construct, this variable was left in the analysis, as its removal would violate structure (Jarvis, MacKenzie, & Podsakoff, Reference Jarvis, MacKenzie and Podsakoff2003). Results support grouping the variables for adequate regression analyses.

Table 2 Pattern matrix

Note. N=181. Principal component analysis with a promax rotation with three factors.

Owing to the unpredicted negative correlation that the cognitive flexibility self-report and external-rater report had to the DV, it was not possible to create a proper formative or composite score for these methods. Therefore, reliability analysis was run to see the α score for factors 2 and 3 established through the factor analysis. The collective group of external-rater reports had a Cronbach’s α of 0.70, which suggests a strong grouping. The self-report measures had an α of 0.50, which is weak yet adequate. The performance measure did not require a reliability score as it will be combined into a composite score or formative construct (Jarvis, MacKenzie, & Podsakoff, Reference Jarvis, MacKenzie and Podsakoff2003).

Tests of hypotheses

Before analysis, the two intelligence scores were formed in a factor score as both are well-validated instruments (Hair, Anderson, Tatham, & Black, Reference Hair, Anderson, Tatham and Black1998). The verbal (crystallized) and spatial measures (fluid) of intelligence when combined, create a general cognitive intelligence factor (Cattell, Reference Cattell1963; Horn, Reference Horn1985).

Hypothesis 1:

Table 3 shows all the regression results. Model 1 introduced the intelligence factor. Results demonstrate a significant positive relationship of the intelligence factor with the adaptive performance outcome on the NFC, supporting Hypothesis 1. The intelligence factor was then used in each of the subsequent models as a control variable.

Table 3 Regression table for cognitive agility measures

Note. N=181.

p<.1; *p<.05; **p<.01.

Hypothesis 2

Hypothesis 2 predicted the formative construct of cognitive agility, as measured by the performance tests, would demonstrate significant variance beyond the intelligence factor. Model 2 shows regression results for cognitive agility, as measured by the performance test beyond intelligence, providing 11% unique variance on the NFC. Hypothesis 2 was supported. This hypothesis was tested in a prior study with an alternative theoretical framing (Good & Michel, Reference Good and Michel2013).

Hypothesis 3

Hypothesis 3 sought to demonstrate that the formative self-report construct (composite of all three self-report measures) of cognitive agility would explain significant variance in adaptive performance beyond intelligence. As mentioned earlier, given the negative correlations that both of the cognitive flexibility scales (self-report and external-rater report) had to the DV, the self-reports and external-rater reports could not be made into a composite or true formative measure; as the negative correlations would dilute the power of the measure. Therefore, as an alternative, linear regressions were used entering all three self-report measures into a single block listwise. This allowed each of the individual measures to interactively demonstrate impact on the DV rather than have the cognitive flexibility measure(s) dilute the results with a negative correlation. Model 3 shows that Hypothesis 3 was upheld as the addition of the self-reports accounted for 6% unique explained variance beyond intelligence in the adaptive performance score. Though a closer examination of the results demonstrate findings that are more nuanced. Focused attention does not add any unique variance (β=0.05, p=.49) and cognitive flexibility has an inverse relationship to the DV (β=−0.19, p=.01).

Hypothesis 4

Hypothesis 4 predicted that the formative construct of cognitive agility measured by external-raters will explain significant variance in the adaptive performance score in the DDM beyond that explained by intelligence. Like the self-reports, a linear regression was used entering all three external-rater measures into a single block listwise as an alternative to a composite score. Model 4 shows that Hypothesis 4 was upheld as the total model explained 6% unique variance on the DV. Like the self-reports, the external-reports have more nuanced results beyond the total change in R 2, including a significant negative association that cognitive flexibility has to the DV (β=−0.32, p<.01).

DISCUSSION

The purpose of this study was to better understand the cognitive capabilities and orientations that support successful adaptation within a real-time dynamic context. While adaptability has become a popular buzzword in management and has been subject to an increasing degree of investigation, individual differences leading to adaptive performance within real-time dynamic tasks remain unclear. Given the increase of DDM demands individual face in organizations, additional investigation is warranted.

As hypothesized, the formative construct of cognitive agility, as assessed by three methods (performance, self-reports and external-rater reports) demonstrated significant variance beyond intelligence in the DDM microworld. Most notable was the impact of cognitive agility as measured by the performance scores (ΔR 2=0.11, p<.001). This finding is meaningful given that previous research using microworlds seldom goes beyond general intelligence in assessing variance of additional variables. In a similarly designed simulation study, both conscientiousness and openness to experience (as measured by the NEO) accounted for 2% unique variance beyond intelligence (LePine, Colquitt, & Erez, Reference LePine, Colquitt and Erez2000). This gives a frame of reference for the potential significance of this finding.

Overall, the performance tests are a more powerful set of predictors in the DDM context than the self- or external-rater reports. This makes theoretical sense based on the implicit nature of the performance tests and the fire chief exercises (Dovidio, Kawakami, & Beach, Reference Dovidio, Kawakami and Beach2001). The self- and external-rater reports are missing this implicit nature, and may have a lower degree of consistent predictability with the DV. Furthermore, each of the performance tests and the DV have a time-intensive component to them; whether it be reaction time (Stroop and go/no go), product creation (in the case of the AUT) or decision speed (with the DV).

Contrary to expectation, negative relationship(s) was found between the questionnaire used for cognitive flexibility and the NFC (self-report β=−0.19, p<.01; external-rater β=−0.32, p<.01). These cognitive flexibility measures were assessed using a metacognitive regulation subscale. Metacognitive regulation accounts for control of cognitive operations leading to greater flexibility (Fernandez-Duque, Baird, & Posner, Reference Fernandez-Duque, Baird and Posner2000). Research shows that metacognitive capacities are linked to expert decision making (Chi, Feltovich, & Glaser, Reference Chi, Feltovich and Glaser1991), successful problem solving (Mayer, Reference Mayer1998) and increased adaptability in uncertain and dynamic contexts (Earley & Ang, Reference Earley and Ang2003). Therefore, it was predicted that metacognitive regulation (i.e., monitoring and control over one’s cognition) would lead to increases in adaptive performance. The results did not support this assertion.

DDM microworlds require a series of continual and real-time rapid decisions, which may represent a completely distinct class of performance, separate from single decisions (Sitkin & Pablo, Reference Sitkin and Pablo1992). Perhaps the ongoing and constant decisions that must be made require one to be less ‘regulated’ in how one thinks. In such a dynamic context, speed of operation compliments accuracy of decisions. Strong metacognitive regulatory tendencies may increase a participant’s likelihood to monitor and evaluate his/her ongoing perception of success and failure (Jacobs & Paris, Reference Jacobs and Paris1987; Schraw & Dennison, Reference Schraw and Dennison1994). Past studies indicate that as one detects error they tend to slow down speed of operation in order to increase accuracy (Rabbit, Reference Rabbit1966; Robertson, Manly, Andrade, Baddeley, & Yiend, Reference Robertson, Manly, Andrade, Baddeley and Yiend1997). Therefore, in this context, participants with strong metacognitive regulation may be more engaged in online monitoring and therefore more aware of possible errors. This awareness may slow down processing at a time when doing so may cause performance decrements. This performance deficit may also be linked to literature on intuitive thought, which suggests that such metacognitive activity creates interference in necessary unconscious material (Kuhn, Reference Kuhn1989; Baylor, Reference Baylor2001). Given the short time horizon of the microworld used here (15 min), slowing down and controlling cognitive operation may have led to a decrease in performance. This may suggest some practical implication for training in which individuals regularly engaged in DDM activity can learn to switch from conscious to automatic processing as real-time dynamism increases (Louis & Sutton, Reference Louis and Sutton1991).

Cognitive agility describes the flexibility between what can be thought of as opposing phenomena (openness and focus). Yet, managing and leading in organizations has been said to require the ability to employ the simultaneous use of opposites (Quinn, Spreitzer, & Hart, Reference Quinn, Spreitzer and Hart1992). Rather than consider them in opposition the goal is to enhance the ‘the ability to act out a cognitively complex strategy by playing multiple, even competing roles, in a highly integrated and complementary way’ (Hooijberg & Quinn, Reference Hooijberg and Quinn1992: 164). Adding to this, organizational ambidexterity (a firm’s ability to manage the dilemma between exploration and exploitation) has been increasingly conceptualized as multi-level phenomena to include an individual-level ambidexterity construct (Mom, van den Bosch, & Volberda, Reference Mom, van den Bosch and Volberda2007). It is likely that there is an individual difference in being able to manage this dilemma well, yet empirical evidence is sparse (Gupta, Smith, & Shalley, Reference Gupta, Smith and Shalley2006). Drawing from a small slice of the data in the current paper, the author has made such a theoretical connection to individual ambidexterity (Good & Michel, Reference Good and Michel2013), suggesting a similarity between being agile and ambidextrous. The more dynamic the context becomes, the more individual ambidexterity is thought to be important for adaptation (Davis, Eisenhardt, & Bingham, Reference Davis, Eisenhardt and Bingham2009). Such dynamism will likely make the management of tension(s) a more important future area of interest for individual difference studies. It may be important to investigate how a capacity like agility can inform other common tensions experienced in organizational life, such as converging/diverging, differentiation/integration and advocacy/inquiry.

Limitations and conclusion

This study has several major limitations that when addressed may provide future research opportunities. This was an initial step in proposing a new construct, and therefore a great deal of future work toward validation is necessary. This could include using a more robust measure of intelligence as a control variable (Salgado, Reference Salgado1999). Future research may consider other variables that could be related to DDM adaptive performance such as working memory (Baddeley & Hitch, Reference Berg and Sternberg1974), situational awareness (Endsely, Reference Endsley1995) and tolerance for ambiguity (Endres, Chowdhury, & Milner, Reference Endres, Chowdhury and Milner2009). The testing scenario of the microworld and the undergraduate sample population substantially limits generalizability of the results to DDM of individuals in organizations. In particular, the choice to use a microworld was as a way to capture the elements found throughout the dynamic experiences individual encounter often within organizational life (Funke, Reference Funke1991; Brehmer & Dorner, Reference Brehmer and Dörner1993; Omodei & Wearing, Reference Omodei and Wearing1995). While few would argue that experiences in organizational life have become more dynamic and complex, the results produced by this microworld-based laboratory study still need to be handled with caution. Results pertaining to cognitive agility and the individual variables which form it (cognitive openness, focused attention and cognitive flexibility) are promising in being able to predict aspects of adaptive performance, yet, this outcome is within the context of a computer-simulated game and laboratory-based decision-making studies do not fully capture real-life decision making (Dawes, Reference Dawes1988).

In conclusion, this study suggests that the formative construct of cognitive agility, as measured by the three performance tests, predicts adaptive performance in the DDM scenario beyond measures of general intelligence. This finding suggests that the unique combination of cognitive openness, focused attention and cognitive flexibility (as they are measured here), may be an important cluster of capabilities in managing real-time dynamic contexts. However, the inverse relationship of the cognitive flexibility questionnaire(s) to the DV raises important questions about the roles of cognitive regulation, cognitive control and cognitive flexibility in real-time DDM contexts. Overall, this study investigates specific capabilities targeted to a specific context. Studies of highly contextualized aspects of adaptive performance are vital to our development in meeting the demands for research and practice in dynamic times.

References

Ackerman, P. L. (1988). Determinants of individual differences during skill acquisition: cognitive abilities and information processing. Journal of Experimental Psychology: General, 117, 288318.Google Scholar
Ackerman, P. L. (1992). Predicting individual differences in complex skill acquisition: Dynamics of ability determinant. Journal of Applied Psychology, 77, 598614.Google Scholar
Anderson, P. (1983). Decision making by objection and the Cuban missile crisis. Administrative Sciences Quarterly, 28, 201222.Google Scholar
Ashford, S. J., & Taylor, M. S. (1990). Adaptation to work transitions: An integrative approach. In G. R. Ferris & K. M. Rowland (Eds.), Research in personnel and human resource management. vol. 8 (pp. 141). Greenwich, CT: JAI Press.Google Scholar
Baddeley, A. D., & Hitch, G. (1974). Working memory. In G. A. Bower (Ed.), The psychology of learning and motivation, vol. 8 (pp. 4789). New York, NY: Academic Press.Google Scholar
Baird, L., & Griffin, D. (2006). Adaptability and responsiveness: The case for dynamic learning. Organization Dynamics, 35, 372383.Google Scholar
Baltes, P. B., & Staudinger, U. M. (1996). Interactive minds: Life-span perspectives on the social foundation of cognition. New York, NY: Cambridge University Press.Google Scholar
Barron, F. (1988). Putting creativity to work. In R. J. Sternberg (Ed.), The nature of creativity (pp. 7698). New York, NY: Cambridge University Press.Google Scholar
Baylor, A. L. (2001). A U-shaped model for the development of intuition by level of expertise. New Ideas in Psychology, 19, 237244.CrossRefGoogle Scholar
Berg, C. A., & Sternberg, R. J. (1985). A triarchic theory of intellectual development during adulthood. Developmental Review, 6, 334370.Google Scholar
Bodner, T. (2000). On the assessment of individual differences in mindful information processing: A thesis (Doctoral dissertation, Harvard University).Google Scholar
Boyatzis, R. E., Goleman, D., & Rhee, K. (1999). Clustering competence in emotional intelligence: Insights from the Emotional Competence Inventory (ECI). In R. Bar-On & J. D. Parker (Eds.), Handbook of emotional intelligence (pp. 343362). San Francisco, CA: Jossey-Bass.Google Scholar
Brehmer, B. (1992). Dynamic decision making: Human control of complex systems. Acta Psychologica, 81, 211241.Google Scholar
Brehmer, B. (1995). Feedback delays in complex dynamic decision tasks. In P. French & J. Funke (Eds.), Complex problem-solving: The European perspective (pp. 103130). Mahwah, NJ: Lawrence Earlbaum Associates.Google Scholar
Brehmer, B., & Dörner, D. B. (1993). Experiments with computer-simulated microworlds: Escaping both the narrow straits of the laboratory and the deep blue sea of the field study. Computers in Human Behavior, 9, 171184.Google Scholar
Brehmer, B., Leplat, J., & Rasmussen, J. (1991). Use of simulation in the study of complex decision making. In J. Rasmussen, B. Brehmer, & J. Leplat (Eds.), Distributed decision making: Cognitive models for co-operative work (pp. 373386). New York, NY: Wiley.Google Scholar
Briscoe, J. P., & Hall, D. T. (1999). Grooming and picking leaders using competency frameworks: Do they work? An alternative approach and new guidelines for practice. Organizational Dynamics, 28, 3752.Google Scholar
Burns, R., & Gallini, J. (1983). The relation of cognitive and affective measures to achievement during an instructional sequence. Instructional Science, 12, 103120.Google Scholar
Cañas, J. J., Quesada, J. F., Antolí, A., & Fajardo, I. (2003). Cognitive flexibility and adaptability to environmental changes in dynamic complex problem solving tasks. Ergonomics, 46, 482501.Google Scholar
Cattell, R. B. (1963). Theory of fluid and crystallized intelligence: A critical experiment. Journal of Educational Psychology, 54, 122.CrossRefGoogle Scholar
Chamorro-Premuzic, T. (2006). Creativity versus conscientiousness: Which is a better predictor of student performance? Applied Cognitive Psychology, 20, 521531.Google Scholar
Chan, D., & Schmitt, N. (2000). Interindividual differences in intraindividual changes in proactivity during organizational entry: A latent growth modeling approach to understanding newcomer adaptation. Journal of Applied Psychology, 85, 190210.Google Scholar
Chi, M. T. H., Feltovich, P. J., & Glaser, R. (1991). Categorization and representation of physics problems by experts and novices. Cognitive Sciences, 5, 121152.Google Scholar
Clark, H. H. (1996). Using language. Cambridge: Cambridge University Press.Google Scholar
Costa, P. T., & McCrae, R. R. (1985). The NEO Personality Inventory manual. Odessa, FL: Psychological Assessment Resources.Google Scholar
Davis, J. P., Eisenhardt, K. M., & Bingham, C. B. (2009). Optimal structure, market dynamism, and the strategy of simple rules. Administrative Science Quarterly, 5, 413452.Google Scholar
Dawes, R. M. (1988). Rational choice in an uncertain world. San Diego, CA: Harcourt Brace Jovanovich.Google Scholar
Derryberry, D., & Rothbart, M. K. (1988). Affect, arousal, and attention as components of temperment. Journal of Personality and Social Psychology, 55, 958966.Google Scholar
DiFonzo, N., Hantula, D. A., & Bordia, P. (1998). Microworlds for experimental research: Having your (control & collection) cake, and realism too. Behavior Research Methods, Instruments, & Computers, 30, 278286.Google Scholar
Digman, J. M. (1990). Personality structure: Emergence of the five-factor model. Annual Review of Psychology, 41, 417440.Google Scholar
Dovidio, J. F., Kawakami, K., & Beach, K. R. (2001). Implicit and explicit attitudes: Examination of the relationship between measures of intergroup bias. In R. Brown & S. L. Gaertner (Eds.), Blackwell handbook of social psychology: Intergroup processes (pp. 175197). Malden, MA: Blackwell.Google Scholar
Dupuy, H. P. (1974). The rationale, development and standardization of a basic vocabulary test. Washington, DC: US Government Printing Office.Google Scholar
Earley, P. C., & Ang, S. (2003). Cultural intelligence: Individual interactions across cultures. Stanford, CA: Stanford University Press.Google Scholar
Edmondson, A. C., Bohmer, R. M., & Pisano, G. P. (2001). Disrupted routines: Team learning and new technology implementation in hospitals. Administrative Science Quarterly, 46, 685716.Google Scholar
Ekstrom, R. B., French, J. W., & Harman, H. H. (1976). Manual for kit of factor referenced cognitive tests. Princeton, NJ: Educational Testing Service.Google Scholar
Elliot, T., Welsh, M., Nettelbeck, T., & Mills, V. (2007). Investigating naturalistic decision making in a simulated micro-world: What questions should we ask? Behavior Research Methods, 39, 901910.Google Scholar
Endres, M. L., Chowdhury, S., & Milner, M. (2009). Ambiguity tolerance and accurate assessment of self-efficacy in a complex decision task. Journal of Management & Organization, 15, 3146.Google Scholar
Endsley, M. R. (1995). Measurement of situation awareness in dynamic systems. Human Factors, 37, 6584.Google Scholar
Fernandez-Duque, D., Baird, J. A., & Posner, M. I. (2000). Executive attention and metacognitive regulation. Consciousness and Cognition, 9, 288307.Google Scholar
Flavell, J. H. (1979). Metacognition and cognitive monitoring: A new area of cognitive development inquiry. American Psychologist, 34, 906911.Google Scholar
Funke, J. (1991). Solving complex problems: Human identification and control of complex systems. In R. J. Sternberg & P. A. Frensch (Eds.), Complex problem solving: Principles and mechanisms (pp. 185222). Hillsdale, NJ: Lawrence Erlbaum Associates.Google Scholar
Glynn, M. A. (1996). Innovative genius: A framework for relating individual and organizational intelligences to innovation. Academy of Management Review, 21, 10811111.CrossRefGoogle Scholar
Gonzalez, C., Vanyukov, P., & Martin, M. K. (2005). The use of microworlds to study dynamic decision making. Computers in Human Behavior, 21, 273286.Google Scholar
Good, D. J., & Michel, E. J. (2013). Individual ambidexterity: Exploring and exploiting in dynamic contexts. Journal of Psychology: Interdisciplinary and Applied, 147, 435453.CrossRefGoogle ScholarPubMed
Gottfredson, L. S. (1997). Intelligence and social policy. Intelligence, 24, 1320.Google Scholar
Gough, H. G. (1979). A creative personality scale for the adjective check list. Journal of Personality and Social Psychology, 37, 13981405.Google Scholar
Guilford, J. P., Christensen, P. R., Merrifield, P. R., & Wilson, R. C. (1978). Alternate uses: Manual of instructions and interpretation. Orange, CA: Sheridan Psychological Services.Google Scholar
Gupta, A. K., Smith, K. G., & Shalley, C. E. (2006). The interplay between exploration and exploitation. Academy of Management Journal, 4, 693706.Google Scholar
Hair, J. F., Anderson, R. E., Tatham, R. L., & Black, W. C. (1998). Multivariate data analysis (4th ed.). New Jersey, USA: Prentice Hall.Google Scholar
Haynie, M. J., Shepherd, D., Mosakowski, E., & Earley, C. P. (2010). A situated metacognitive model of the entrepreneurial mindset. Journal of Business Venturing, 25, 217229.Google Scholar
Hesketh, B. (1997). Dilemmas in training for transfer and retention. Applied Psychology: An International Review, 46, 317319.Google Scholar
Hooijberg, R., & Quinn, R. E. (1992). Behavioral complexity and the development of effective managers. In R. L. Phillips & J. G. Hunt (Eds.), Strategic leadership: A multiorganizational-level perspective (pp. 161176). Westport, CT: Quorum Books.Google Scholar
Horn, J. L. (1985). Remodeling old models of intelligence. In B. B. Wolman (Ed.), Handbook of intelligence: Theories, measurements, and applications (pp. 267300). New York, NY: Wiley.Google Scholar
Hunter, J. E., & Hunter, R. F. (1984). Validity and utility of alternative predictors of job performance. Psychological Bulletin, 76, 7293.Google Scholar
Isen, A. M., Daubman, K. A., & Nowicki, G. P. (1987). Positive affect facilitates creative problem solving. Journal of Personality and Social Psychology, 52, 11221131.Google Scholar
Jacobs, J. E., & Paris, S. G. (1987). Children’s metacognition about reading: Issues in definition, measurement, and instruction. Educational Psychologist, 22, 255278.Google Scholar
Jarvis, C., MacKenzie, S., & Podsakoff, P. (2003). A critical review of construct indicators and measurement model misspecification in marketing and consumer research. Journal of Consumer Research, 30, 199218.Google Scholar
Keilp, J. G., Sackeim, H. A., & Mann, J. J. (2005). Correlates of trait impulsiveness in performance measures and neuropsychological tests. Psychiatry Research, 135, 191201.Google Scholar
Koch, C. (2003). Self-monitoring, need for cognition, and the Stroop effect: A preliminary study. Perceptual and Motor Skills, 96, 212214.Google Scholar
Kozlowski, S. W. J., Gully, S. M., Brown, K. G., Salas, E., Smith, E. M., & Nason, E. R. (2001). Effects of training goals and goal orientation traits on multidimensional training outcomes and performance adaptability. Organizational Behavior & Human Decision Processes, 85, 131.Google Scholar
Kuhl, J., & Kazen-Saad, M. (1988). A motivational approach to volition: Activation and de-activation of memory representations related to unfulfilled intentions. In V. Hamilton, G. H. Bower, & N. H. Firjda (Eds.), Cognitive perspectives on emotion and motivation (pp. 6385). Dordrecht, The Netherlands: Martinus Nijhoff.Google Scholar
Kuhn, D. (1989). Children and adults as intuitive scientists. Psychological Review, 96, 674689.Google Scholar
Langer, E. (1989). Minding matters: The consequences of mindlessness-mindfulness In L. Berkowitz (Ed.), Advances in experimental social psychology (pp. 137173). San Diego, CA: Academic Press.Google Scholar
LePine, J. A., Colquitt, J. A., & Erez, A. (2000). Adaptability to changing task contexts: Effects of general cognitive ability, conscientiousness, and openness to experience. Personnel Psychology, 53, 563593.Google Scholar
Lerch, F. J., & Harter, D. E. (2001). Cognitive support for real-time dynamic decision making. Information Systems Research, 12, 6382.Google Scholar
Littman, J. A. (2005). Curiosity and the pleasures of learning: Wanting and liking new information. Cognition and Emotion, 19, 793814.Google Scholar
Louis, M. R., & Sutton, R. I. (1991). Switching cognitive gears: From habits of mind to active thinking. Human Relations, 44, 5576.Google Scholar
Luchins, A. S., & Luchins, E. H. (1959). Rigidity in behavior. Eugene, OR: University of Oregon Press.Google Scholar
Lustig, C., May, C. P., & Hasher, L. (2001). Working memory span and the role of proactive interference. Journal of Experimental Psychology General, 130, 199207.Google Scholar
MacLeod, C. M. (1991). Half a century of research on the Stroop effect: An integrative review. Psychological Bulletin, 109, 163203.Google Scholar
Mainemelis, C., Boyatzis, R., & Kolb, D. A. (2002). Learning styles and adaptive flexibility: Testing experiential learning theory. Management Learning, 33, 533.Google Scholar
Martin, M. M., & Anderson, C. M. (1998). The cognitive flexibility scale: Three validity studies. Communication Reports, 11, 19.Google Scholar
Martin, M. M., & Rubin, R. B. (1995). A new measure of cognitive flexibility. Psychological Reports, 76, 623626.Google Scholar
Martindale, C., Anderson, K., Moore, K., & West, A. N. (1996). Creativity, oversensitivity, and rate of habituation. Personality and Individual Differences, 20, 423427.Google Scholar
Matthews, G., & Deary, I. J. (1998). Personality traits. Cambridge, UK: Cambridge.Google Scholar
Mayer, R. (1998). Cognitive, metacognitive, and motivational aspects of problem solving. Instructional Science, 26, 4963.Google Scholar
Mitroff, S. R., Simons, D. J., & Franconeri, S. L. (2002). The siren song of implicit change detection. Journal of Experimental Psychology: Human Perception and Performance, 28, 798815.Google Scholar
Mom, T. J., van den Bosch, F. J., & Volberda, H. W. (2007). Investigating managers’ exploration and exploitation activities: The influence of top-down, bottom-up, and horizontal knowledge inflows. Journal of Management Studies, 44, 910931.Google Scholar
Morrison, R. (1977). Career adaptivity: The effective adaptation of managers to changing role demands. Journal of Applied Psychology, 62, 549558.Google Scholar
Moses, L. J., & Baird, J. A. (1999). Metacognition. The MIT Encyclopedia of the Cognitive Sciences, 533535.Google Scholar
Mumford, M. D., Zaccaro, S. J., Harding, F. D., Jacobs, T. O., & Fleishman, E. A. (2000). Leadership skills for a changing world: Solving complex social problems. Leadership Quarterly, 11, 1135.Google Scholar
Neisser, U. (1967). Cognitive psychology. New York, NY: Appleton-Century-Crofts.Google Scholar
Omodei, M. M., & Wearing, A. J. (1995). The fire chief microworld generating program: An illustration of computer-simulated microworlds as an experimental paradigm for studying complex decision-making behavior. Behavior Research Methods, Instruments & Computers, 27, 303316.Google Scholar
Omodei, M. M., Wearing, A. J., McLennan, J., Hansen, J., Clancy, J. M., Elliott, G. C., Ley, T., Taranto, P., & Thorsteinsson, E. B. (2001). Human decision making in complex systems interim summary report: Research agreement #2, unpublished manuscript Melbourne: La Trobe University.Google Scholar
Pallier, G., Roberts, R. D., & Stankov, L. (2000). Biological versus psychometric intelligence: Halstead’s (1947) distinction re-visited. Archives of Clinical Neuropsychology, 13, 205226.Google Scholar
Pintrich, P. R., Smith, D. A. F., Garcia, T., & McKeachie, W. J. (1991). A manual for the use of the Motivated Strategies for Learning Questionnaire (MSLQ). Report 91-B-004. Ann Arbor, MI: National Center for Research to Improve Postsecondary Teaching and Learning. 87pp.Google Scholar
Pulakos, E. D., Arad, S., Donovan, M. A., & Plamondon, K. E. (2000). Adaptability in the workplace: Development of a taxonomy of adaptive performance. Journal of Applied Psychology, 85, 612624.Google Scholar
Quinn, R. E., Spreitzer, G. M., & Hart, S. (1992). Challenging the assumptions of bipolarity: Interpenetration and effectiveness. In S. Srivastva & R. Fry (Eds.), Executive continuity (pp. 222252). San Francisco, CA: Jossey-Bass.Google Scholar
Rabbit, P. M. A. (1966). Errors and error correction in choice reaction tasks. Journal of Experimental Psychology, 71, 264272.Google Scholar
Reder, L. M., & Schunn, C. D. (1996). Metacognition does not imply awareness: Strategy choice is governed by implicit learning and memory. In L. M. Reder (Ed.), Implicit memory and metacognition (pp. 4578). Hillsdale, NJ: Erlbaum.Google Scholar
Rende, B. (2000). Cognitive flexibility: Theory, assessment, and treatment. Seminars in Speech and Language, 21, 121133.Google Scholar
Rigas, G., Carling, E., & Brehmer, B. (2002). Reliability and validity of performance measures in microworlds. Intelligence, 30, 463480.Google Scholar
Robertson, I. H., Manly, T., Andrade, J., Baddeley, B. T., & Yiend, J. (1997). Oops!: Performance correlates of everyday attentional failures in traumatic brain injured and normal subjects. Neuropsychologia, 35, 747758.Google Scholar
Rossiter, J. R. (2002). The C-OAR-SE procedure for scale development in marketing. International Journal of Research in Marketing, 19, 305335.Google Scholar
Salgado, J. F. (1999). Personnel selection methods. International Review of Industrial an Organizational Psychology, 14, 154.Google Scholar
Schraw, G., & Dennison, R. S. (1994). Assessing metacognitive awareness. Contemporary Educational Psychology, 19, 460475.CrossRefGoogle Scholar
Shapiro, K. L., & Raymond, J. E. (1994). Temporal allocation of visual attention: Inhibition or interference. In D. Dagenbach & T. Carr (Eds.), Inhibitory process in attention, memory and language (pp. 151188). New York, NY: Academic Press.Google Scholar
Shanteau, J. (1988). Psychological characteristics and strategies of expert decision makers. Acta Psychologica, 68, 203215.CrossRefGoogle Scholar
Sitkin, S. B., & Pablo, A. L. (1992). Reconceptualizing the determinants of risk behavior. Academy of Management Review, 17, 938.Google Scholar
Spiro, R. J., & Jehng, J. C. (1990). Cognitive flexibility and hypertext: Theory and technology for the nonlinear and multi-dimensional traversal of complex subject matter. In D. Nix & R. J. Spiro (Eds.), Cognition, education and multi-media: Exploring ideas in high technology, Chapter 7 (pp. 163205). Hillsdale, NJ: Lawrence Erlbaum Associates.Google Scholar
Sternberg, R. J. (1997). Thinking styles. New York, NY: Cambridge University Press.Google Scholar
Sternberg, R. J. (1999). The theory of successful intelligence. Review of General Psychology, 3, 292316.Google Scholar
Stroop, J. R. (1935). Studies of interference in serial verbal reactions. Journal of Experimental Psychology, 12, 643662.Google Scholar
Theeuwes, J. (1994). Endogenous and exogenous control of visual selection. Perception, 23, 429440.Google Scholar
Tversky, A., Sattath, S., & Slovic, P. (1988). Contingent weighting in judgment and choice. Psychological Review, 95, 371384.Google Scholar
Wallach, M. A., & Kogan, N. (1965). Modes of thinking in young children. New York: Holt, Rinehart & Winston.Google Scholar
Yantis, S. (1993). Stimulus-driven attentional capture. Current Directions in Psychological Science, 2, 156161.Google Scholar
Zaccaro, S. J. (2001). The nature of executive leadership: A conceptual and empirical analysis of success. Washington, DC: APA Books.Google Scholar
Zimmermann, P., & Fimm, B. (2000). Test for attentional performance (TAP). Herzogenrath: PSYTEST.Google Scholar
Figure 0

Figure 1 Formative construct of cognitive agility

Figure 1

Figure 2 Networked Fire Chief

Figure 2

Table 1 Descriptive statistics

Figure 3

Table 2 Pattern matrix

Figure 4

Table 3 Regression table for cognitive agility measures