skip to main content
research-article
Open access

Naturalistic Digital Behavior Predicts Cognitive Abilities

Published: 30 August 2024 Publication History

Abstract

Individuals are known to differ in cognitive abilities, affecting their behavior and information processing in digital environments. However, we have a limited understanding of which behaviors are affected, how, and whether some features extracted from digital behavior can predict cognitive abilities. Consequently, researchers may miss opportunities to design and support individuals with personalized experiences and detect those who may benefit from additional interventions. To characterize digital behaviors, we collected 24/7 screen recordings, input behavior, and operating system data from the laptops of 20 adults for two weeks. We use cognitive test results from the same individuals to characterize their cognitive abilities: psychomotor speed, processing speed, selective attention, working memory, and fluid intelligence. Our results from regression analysis, path modeling, and machine learning experiments show that cognitive abilities are associated with differences in digital behavior and that naturalistic behavioral data can predict the cognitive abilities of individuals with small error rates. Our findings suggest naturalistic interaction data as a novel source for modeling cognitive differences.

1 Introduction

Differences in cognitive ability impact our everyday digital behavior by constraining how we perceive, process, and act in the digital world. For example, individuals with reduced cognitive ability might tend to read a document longer than individuals with higher cognitive ability; they might require more time to evaluate information [9] and need additional effort to navigate digital spaces to locate important information [60]. However, current “one-size-fits-all” systems do not account for users with varying cognitive abilities, resulting in higher cognitive loads, and frustration for individuals with lower cognitive abilities [60]. Consequently, understanding cognitive differences in digital behavior is important for designing user interfaces that are tailored to individuals’ cognitive abilities. For instance, reducing the complexity of user interfaces by adjusting the amount of information presented to match a user’s cognitive ability; reducing an overwhelming number of documents retrieved and presented to the user on a search interface or a recommender system; or selecting assistive input options to meet users’ restricted input abilities.
Research in the intersection of cognitive science and human-computer interaction (HCL) has studied the relationship between human digital behavior and cognitive abilities [33, 53]. It has shown that limitations in cognitive abilities can significantly impact how individuals interact with digital systems on many levels, such as imposing extra difficulties in their sensory input, information processing, and motor control [32]. Sensory input abilities may limit how individuals recognize and perceive information from their environment [32], which could hamper their interactions with computer systems [62]. Information processing abilities can limit how fast users interpret presented information and can respond to it [9, 27, 32]. This includes the amount of time required for users to make click decisions [8], the time users dwell on Web pages to obtain information [9], and the time spent on browser tabs [13]. Motor control abilities can affect how fast and accurately users can operate user interfaces and input devices [3]. In particular, users having reduced cognitive ability tend to experience more difficulties in performing complex interactions involving multiple keystrokes [3, 73]. Together information perception, information processing, and motor control are interwined to determine the overall digital behavior of an individual.
In summary, digital behavior involves several cognitive processes, which can be impacted by an individual’s cognitive abilities. Limits by cognitive abilities may lead to difficulties in perceiving, processing, and executing tasks with a computer system. These include declines in: (1) the ability to maintain focus on relevant information (selective attention) while switching applications [10, 25]; (2) the ability to process and understand the visual layout of a Website or search engine result pages (processing speed) [10]; (3) the ability to reason (fluid intelligence) [33], figuring out relevant information and links to use for task completion; (4) the ability to recall information for successful Web navigation (working memory) [13]; and (5) the ability to react to useful information (psychomotor speed) and perform motor control during interactions [25].
However, previous studies have focused on lab-based experiments conducted in controlled environments and experimental designs [8, 10, 33]. The advantage of laboratory settings is that task difficulty, including time limits, knowledge levels, or expected physical efforts, can be manipulated so that cognitive differences can be revealed [33, 46, 61]. The disadvantage of laboratory settings, on the other hand, is that the observed interaction data may not be representative of the user’s naturalistic real-world behavior [29, 45, 61, 67]. As a result, researchers and developers may have missed opportunities to detect cognitive differences from naturalistic digital behavior for implementing systems and services that can account for people with varying cognitive abilities.
Recent research has also aimed at building models to predict the difficulty of various cognitive tasks. Steichen et al. [52] and Conati et al. [13] used eye tracking to infer users’ cognitive abilities in visualization inspection tasks. Kangasrääsiö et al. [31] built models for predicting future behavior for skill acquisition and visual search. Gordon et al. [25] used smartphone application usage behavior to predict the cognitive ability of the users. These results mark a trend toward models that use behavioral traits to assess and predict cognitive ability. However, previous work has been limited to scoped tasks, such as gaze tracking on visualizations in laboratory settings or application usage patterns with limited interaction data.
Here, we study how cognitive abilities are associated with everyday digital behaviors. We examine the use of naturalistic behavioral data sourced from in-the-wild monitoring of computer usage to predict cognitive abilities. In particular, we seek answers to the following research questions (RQs):
RQ1: How is cognitive ability associated with naturalistic digital behaviors?
RQ2: Can we predict cognitive abilities from naturalistic digital behaviors?
To answer the RQs, we present a study recording and analyzing 20 individuals’ digital behaviors and cognitive abilities. The overall procedure of our study is as follows and illustrated in Figure 1. Digital activity monitoring software was installed on the participants’ computers for 14 days, recording participants’ digital behavior 24/7. A dataset consisting of screen recordings, OS logs, keystrokes, and mouse inputs of participants was then used to associate cognitive abilities with various digital behaviors: information perception (measured with cumulative text and cumulative entropy of information on the screen); information processing (measured with the number of tabs, number of pages visited, and number of Websites visited); and input behavior (measured as typing speed on the text and search input elements). Each participant took part in cognitive testing on five cognitive abilities: psychomotor speed, processing speed, selective attention, working memory, and fluid intelligence. To explore how differences in cognitive ability might explain differences we observe in digital behaviors, we conducted experiments using digital behaviors paired with the cognitive testing results of the participants. We generalized these results into a predictive model capable of predicting cognitive abilities from digital behaviors. To our knowledge, our work is the first to exploit rich interaction logs from naturalistic everyday computer usage behavior to study and predict cognitive ability.
Fig. 1.
Fig. 1. The overall procedure of the user study. First, digital activity monitoring software was installed on the participants’ laptops for 14 days to collect screen content, operating system (OS) logs, and keystrokes. Second, digital behaviors were extracted from the digital monitoring data, including information perception, information processing, and input behavior. Third, participants performed a battery of cognitive tests that were used to quantify their cognitive abilities. Fourth, the effects of cognitive abilities on digital behaviors were analyzed. Fifth, predictive models were built to predict cognitive abilities.
The results of the experiments revealed the following:
Higher selective attention is associated with an increased amount of information processed from multiple sources. For example, users with higher selective attention change tabs more often than those with lower selective attention.
Higher fluid intelligence is associated with faster information processing. For example, users with higher fluid intelligence processed information by spending less time scanning information than those with lower fluid intelligence scores.
Higher working memory capacity is associated with increased speed of writing. For example, users with higher working memory capacity typed text and formulated queries faster.
Machine learning models demonstrate the possibility of predicting cognitive abilities with a small error and significantly outperform the control models.

2 Background

Our work is based on studies of behavioral and cognitive differences in HCI, which are reviewed below.

2.1 Behavioral Differences

Individual differences in digital behavior have been extensively studied [9, 14, 39, 54]. A series of studies (Table 1) indicated that individuals exhibit varying behaviors in digital environments. For instance, prior works investigated how users processed information on the Web pages, such as how long the users dwell on a Web page [38], how many tabs the users opened, and how long these tabs were kept open [22]. Studies by Dubroy and Balakrishnan [22] revealed that a specific group of users switched tabs and changed pages more often than others. Similarly, the duration for which users spend on a page or Website, also known as dwelling time, has been used to reveal underlying cognitive processing in language comprehension and decision making [38]. Navalpakkam and Churchill [38] found that dwelling time and mouse tracking can help measure user attention on Web pages. However, their studies did not consider cognitive abilities associated with behavioral differences. Another line of research has focused on predicting the Web page the users would open next [11, 42]. They found that users with similar behavior patterns belong to the same group and would visit the same Web pages [42]. Although their studies were limited in context, they found that various behavioral patterns could emerge across user groups.
Table 1.
ResearchReferencesBehavioral dataBehavioral factors / cognitive abilitiesData sources
Behavioral differencesLut et al. [35]Browser usage, backButton, Omnibox (address bar/search box), TabChange, urlChangeInformation processingWeb browsing
Gordon et al. [25]App use time, umber of app launches, text messagesInformation processingSmartphone logs, web browsing
Kumar and Tomkins [34]Number of pageviews, amount of usageInformation processingWeb browsing
Dubroy and Balakrishnan [22]Number of tabs, revisitationInformation processingWeb browsing
Crichton et al. [17]Tab changes per hour, power, URLs, pages visited per day, search, revisitation rateInformation processing, input behaviorWeb browsing other apps
Tossell et al. [55]URL visits, page visits, browsing duration, query rate, site revisitation, page revisitationInformation perception, information processing, input behaviorWeb browsing, maps, weather, news, emails
Cao et al. [5]Email and file usage time, remote meeting durationInformation processingWeb, email, productivity-related files
Chen et al. [7]Gaze time for traveling between visual cues, dwell duration, the time for extracting information from cues.Information perception, input behaviorCustomized user interface, eye track
Jokinen et al. [30]Visits to already visited visual locationsInformation perceptionNative OS interface eye track
Kangasrääsiö et al. [31]Visual information, keypress, menu selectionsInformation perception, input behaviorCustomized user interfaces, eye track
Connecting cognition to individual differencesKaranam and van Oostendorp [33]Number of Web page visits, number of query reformulation, task completion timeFluid intelligenceSearch behavior
Chin et al. [10]Tab switches, number of Web page visitsProcessing speed, working memory, selective attentionHealth literacy, search behavior
Sharit et al. [50]task completion timePsychomotor speed, selective attention, working memoryWeb browsing, search behavior
Trewin et al. [56]Number of Web page visits, task completion timeFluid intelligenceSearch behavior
Conati et al. [13]Number of tabs, gaze time on visual information, number of clicksWorking memoryWeb browsing
Gordon et al. [25]Number of apps, app use duration, switching speedPsychomotor speed, working memory, attention abilityApp usage, web browsing
Table 1. Overview of Previous Works on Behavioral Differences and Associating Cognitive Abilities to Individual Differences
The way information is presented on a user interface has also been found to affect users’ perception and behavior [8, 9]. Chin and Fu [9] examined how users with lower cognitive performance perceived information presented in health information Websites. They asked users to interact with two user interfaces having different information organizations—one with a higher amount of text focused on health symptoms, and the other with more general information about body parts and less text. They found that lower cognitive performing users, who have less knowledge about health symptoms, tended to spend more time examining individual Web page’s information and made fewer clicks on links within the Web page. On the other hand, higher cognitive performing users spent less time on individual Web pages and made more clicks on links within the page. This was likely due to cognitive differences between individuals, which can impact how users perceive and interact with information in user interface [9, 62]. Zhou et al. [73] also found that lower cognitive performing users struggle with understanding the meaning behind graphic icons, leading to difficulties in using the computer. Furthermore, visual search behavior has been found to be associated with cognition and individual personality traits [19, 40]. For instance, introverts completed visual search tasks longer than extroverts [19]. Additionally, previous studies have suggested that the speed with which individuals understand information, as captured by rapid eye gazes during text skimming, is influenced by cognitive ability [23, 37].
Differences have also been shown in input behavior across individuals [15]. Aula et al. [2] found the expertise level of computer use impacted the speed of typing a search query in the query box. Expert users would formulate queries and evaluate the search results more quickly than novice users. Imbir et al. [28] examined individual differences in writing behavior and such behavioral data was used to infer the user’s emotion during interaction with a word processor [28].
Overall, digital behaviors, including visual information perception, information processing, and typing speed, are well-established factors and have been considered broadly for research studying individual differences [17, 25, 35, 55]. Despite extensive research, to our knowledge, there has been no work on predicting cognitive abilities from naturalistic interaction data. Most previous studies have been based on lab-based experimentation, and results and findings remained hypothetical approximations of real-life digital behavior. For instance, in lab-based studies, users were asked to engage in a simulated work task (e.g., interact with visualizations [7] or perform a search task [10]). Task difficulties were often determined before the study so individual differences could occur [10]. Our research contributes further evidence on the impact of cognitive abilities to users’ online behavior by capturing real-life data through continuous 24/7 monitoring of users’ digital activities in natural settings. To characterize naturalistic digital behavior, we extract features from monitoring data, including cumulative text, cumulative entropy, tab change rate, number of page visits, number of site visits, and typing speed. We consider these general behavioral features as they do not depend on any specific task but on the direct behavior of the users in the digital world. Furthermore, previous studies have typically relied upon data from adults having normal cognitive abilities [17, 22]. In this article, we analyzed the digital behaviors of a diverse sample of users, including both adults with reduced cognitive ability and those with normal cognitive ability.

2.2 Connecting Cognition to Individual Differences

The role of cognition in technology usage has suggested that behavioral differences are linked to users’ cognitive abilities [13, 54, 58, 71] (see Table 1). However, it is unclear whether cognitive abilities are directly linked to digital behavior in the wild. Researchers have studied technology usage and cognitive abilities in specific tasks. Findings from an early study [58] suggested that cognition is an important predictor of technology usage. For instance, increased usage of video games and navigation applications was associated with higher cognitive abilities. In another study, adults with reduced cognitive ability were found to perceive technology as more difficult to use than those without cognitive deficits [47]. However, most studies relied primarily on surveys of older adults and did not consider real-world behavior. A few studies, however, analyzed the actual usage data of the users [16, 33, 51]. Karanam and van Oostendorp [33] considered searching behavior and found that users with higher cognitive abilities completed tasks faster. Crabb and Hanson [16] analyzed the users’ browsing histories and suggested that individuals with higher processing capacity tended to browse more Web pages than their lower-cognitive-ability counterparts. Smith and Chaparro [51] focused on input behavior and found that adults having reduced cognitive ability wrote slower than those without cognitive decline. However, users were instructed to work on an information task in a single uninterrupted lab session, e.g., usually an hour. In real life, users are unlikely to concentrate on a search task for such a long session, making user behavior in the lab different from that in the wild. In addition, those studies mainly focused on the impact of age-related differences in searching and browsing but did not examine how the user’s cognitive ability is associated with different digital behavioral patterns. Age differences have been shown to affect psychomotor and processing speed in Web browsing [10] and restrain user behavior in general [31]. Growing evidence has also been reported on variance in visual scanning being dependent on the cognitive properties of each individual [14, 36].
Working memory is another cognitive ability that has been found to be associated with browsing behavior [10]. Peterson et al. [41] showed how working memory imposed significant constraints on how people work on interactive tasks. Working memory has also been considered in the design of interaction in adaptive systems [30], and information-intensive tasks are performed with strategies that are compliant with individuals’ particular memory limits [54].
Moreover, selective attention ability has been found to affect people’s process of the information they visually attend to [10]. Specifically, users having higher cognitive ability have been shown to perform better in tasks requiring divided attention, such as web and folder navigation tasks [9]. The structure in which the Websites are organized may affect the amount of attention required for the navigation process. The deeper in the Website hierarchy the target page is located, the more attention is needed for the navigation process. Chin and Fu [9] suggested that recommendation systems could minimize such processing demands and better support users having reduced cognitive ability. However, previous research has not focused on predicting cognitive ability, and studies have been conducted in limited contexts, such as navigation tasks within a health information system [9].
Another cognitive ability affecting digital behavior is fluid intelligence. Trewin et al. [56] explored individual differences due to fluid intelligence in information search tasks. The results suggest that individuals with higher fluid intelligence performed computer activities more intensively than those with weaker fluid intelligence.
Research has also aimed beyond finding associations between cognition and behavior by building predictive models to estimate cognitive ability. Predicting cognitive ability has previously been found to be important for personalization for a wide array of services in technology usage [18] or office productivity [26], and in avoiding motor vehicle accidents [70]. Gordon et al. [25] found that users with lower levels of cognitive ability used fewer smartphone apps, and sent few messages, and such behavioral information can be used to predict the cognitive test performance of the users. However, their work focused on a limited task context, such as classifying whether a user would fall into the cognitively young or cognitively old groups but did not directly predict cognitive abilities. Conati et al. [13] considered eye-tracking data to predict the user’s cognitive ability while interacting with visualization. However, their work was conducted in the laboratory and relied solely on eye-tracking data.
Here, we focus on understanding whether the user’s cognitive abilities can be associated with differences in digital behaviors and examine whether it is possible to predict cognitive ability variables from data recorded from real-world naturalistic digital behavior. Although there is already solid evidence that cognitive abilities are related to several behavioral factors [25], most studies examining digital behavior have focused on only a few of these factors in isolation and in controlled laboratory studies. Our unique contribution is to assess the relationships among many cognitive abilities (psychomotor speed, processing speed, selective attention, working memory, and fluid intelligence) with behavioral factors as they occur in the naturalistic in-the-wild behavior of a diverse sample of users. Moreover, we study whether naturalistic behavioral data is predictive of cognitive abilities. That is, whether it is possible to record implicit behavioral logs and use that data to draw reliable predictions on different cognitive abilities.

3 Data Collection Methodology

We conducted an in-the-wild study to collect naturalistic behaviors from everyday digital activities and corresponding cognitive performances.

3.1 Apparatus

We used a digital activity monitoring and screen recording system to record all user behaviors on the computer continuously. The system recorded screen frames at two-second intervals, and OS log information associated with screen frames, including the titles of active windows, the names of active applications, the Uniform Resource Locators (URLs) of Web pages on active applications, and timestamps. In addition, the system also collected keystrokes and mouse behaviors, including clicks and scrolls and associated timestamps. The digital activity monitoring and screen monitoring system was developed for Microsoft Windows OS. We used Desktop App UI to implement the monitoring system. The system performed monitoring functions: saving active windows as images and collecting the aforementioned OS log information. In addition, a stopping function was also implemented to allow the users to stop the monitoring.

3.2 Participants

To ensure our sample of participants reflects variance in cognitive ability, we recruited participants with varying ages as age is often associated with cognitive differences. Overall, 20 individuals aged 19–71 participated in our study (Table 2). Ten reported themselves as males, ten as females, and none reported themselves as other gender. Their mean age was 40.4, SD 18.3. It should be noted that we did not use age grouping in the study, but the purpose was to have representative samples of varying aged users. This design choice allowed us to acquire participants with a wide range of cognitive abilities and observe behavior differences among them.
Table 2.
 P1P2P3P4P5P6P7P8P9P10P11P12P13P14P15P16P17P18P19P20
GenderMFFFMMFMFFMMMFMMFMFF
Age1919222427303035355051525457596062636371
Table 2. Participants’ Gender (Self-Reported) and Age Information
F, Female; M, Male.
The participants were recruited via a posting that was distributed to the University of Helsinki’s mailing lists and alumni communities. A questionnaire (see Appendix A.1) was attached to the recruiting message to collect background information on potential candidates. Only respondents who used laptops as their main devices for everyday digital activities and had a higher education background were considered eligible for the study. Another eligibility criterion for participating in the study was that respondents were required to have sufficient skills and experience in using various software. Respondents had to answer nine open-ended questions assessing their theoretical and practical knowledge in the context of the usage of computers and software applications. Respondents were also asked to work on a software installation task. They were asked to follow a guide for installing a specific software application on their computers. Only respondents who answered all questions correctly and managed to install the software were qualified for the study. All participants had laptops of recent models and had Windows 10 OS installed on their laptops. Such criteria of eligibility would help reduce the potential confounding effects of other factors on the overall results of the experiment. For instance, users with lower levels of education and computer knowledge might have less interest in using their computers. This could affect their computer use in frequency, intensity, and duration. They might not use their computer as much, or for as long, as other users with a higher level of education and computer knowledge, which would have become a factor in the experiment.
Participants were all knowledge workers. They were University researchers, faculty members, students, and city administration workers. Out of the 20 participants, six participants had a graduate degree as their highest level of education, 12 participants had an undergraduate degree, and two participants were in undergraduate programs. All participants had advanced computer skills, were experienced computer users, and their working language was English. All participants were healthy, right-handed, and with normal or corrected-to-normal vision.
The participants were informed of our privacy guidelines prior to joining the experiment. They were told that the monitoring of behavioral data was stored on their computers during the monitoring phase. Afterward, the data would be transferred to a secured server and used only for research purposes. After the experiment, the participants were compensated with 150 Euros. A consent form was obtained from the participants regarding the data usage, privacy, and experiment procedure.

3.3 Digital Activity Monitoring

Upon agreeing to participate in the experiment, the digital activity monitoring and screen recording system was installed on each participant’s laptop and set to run continuously in stealth mode for 14 days. Participants were explained that the monitoring system was automatically launched whenever the laptop was turned on. The whole session including receiving instructions and installing the monitoring system took approximately 15 minutes. The participants were advised to use their laptops as usual and to avoid stopping the monitoring unless necessary during the monitoring phase. After 14 days, the participants visited our lab, and the digital activity monitoring system was uninstalled from their laptops. The monitoring data was then transferred to our secured system. The whole process took approximately 10 minutes. The research was approved by the ethical committee of the University of Helsinki and complied with the declaration for managing data obtained from human participants.

3.4 Digital Activity Data

Table 3 presents the summary statistics of digital activity data. Overall, 1,442,447 (\(M=75,918,\text{SD}=75,570\)) screen frames were recorded from the participants for a duration of 14 days. Participants spent an average of 38.13 (\(\text{SD}=24.41\)) hours over two weeks using the computer.
Table 3.
 MinMaxMeanSD
Number of screen frames15,244237,86475,91875,570
Duration of computer use (in hour)10.5174.4738.1324.41
Table 3. Digital Activity Data of 20 Participants Recorded for a Duration of 14 Days
We examined the data gathered from the participants to ensure that it was approximately balanced across participants with respect to the amount of usage by the hour of the day and by application type. The amount of usage (time spent using the computer) was considered “screen time” during which there were input and output signals from the keyboard, mouse, and information changes on the screen. We computed average screen time by hourly rate. Furthermore, we extracted the average share in percentage for individual applications and categorized them into appropriate categories (Appendix A.3) based on the coding structure from the prior work [35].
Figure 2(a) and 2(b) present participants’ average screen time and average share in percentage for various applications. Participants had balanced screen time and actively used their computers between 8–11 and 13–17. Participants mostly spent time using Web browsers with an average screen time of approximately 80%.
Fig. 2.
Fig. 2. Participants’ average time spent using a computer and applications.

3.5 Digital Behavior Extraction

Table 4 presents digital behavior factors used in the study. Three behavioral factors were studied: information perception, information processing, and input behavior. Information perception refers to how much information a user perceives by measuring the amount of cumulative text and the number of cumulative entropy users consumed. Information processing refers to how much information a user processes by measuring the number of tab changes, page visits, and site visits. These measures were normalized to be on an hourly rate as participants may spend a varying amount of time on their computers. Lastly, input behavior describes how fast a user provides input to a computer system by key presses. We measured input behavior by computing the average typing speed of texts entered in the Omnibox (address bars and search boxes). Behavioral factors were extracted from the digital activity data of 20 participants and are described in greater detail below:
Cumulative text per hour: Text was extracted from screenshots. The accumulation of new words was calculated as the sum of the new words compared to the previous screenshot across all screenshots and normalized to represent an average hourly rate.
Cumulative entropy per hour: We followed prior research to compute the entropy [43, 49]. The approach was to capture the level of details of the screenshots to understand the processing capability of an individual. Screenshots were converted from color to grayscale. The entropy was computed by summing the proportion of pixels with each grayscale color weighted by its logarithm. We used a screen difference solution to only consider new pixels between the screens. The sum of all screenshots’ entropies was normalized for an average hourly rate to compute cumulative entropy.
Tab change per hour: The number of tab switches was counted and normalized for an average hourly rate. Tab switches included alternating between tabs that were already open.
Page visit per hour: The number of unique URLs opened in any Web browser and normalized for an average hourly rate.
Site visit per hour: The count of unique Web domains opened and normalized for an average hourly rate.
Omnibox typing speed: Typing events (key-press) occurring during the focus on an address bar of a Web browsing software or a search engine’s query box. For Omnibox typing speed, we averaged the duration (in milliseconds) for a user to type a single character.
Table 4.
Behavioral FactorsDefinitionsMeasuresReferences
Information perceptionThe amount of information appearing on the screen that users could perceive.Cumulative text per hourThe number of cumulative new words appear on the screen [19]
 Cumulative entropy per hourThe level of detail of the information on the screen, higher entropy indicates more details and complex information content [49, 72]
Information processingThe number of information sources that the users could attend to.Tab change per hourAlternating between existing tabs [17, 35]
Page visits per hourThe number of unique Web pages visited [17, 55]
Site visits per hourThe number of unique Website visited [17, 55]
Input behaviorInteraction with the system.Omnibox typing speed (the amount of time in milliseconds taken to press a key)Typing speed in the address bar and search engine [35]
Table 4. Description of Digital Behavioral Factors
These include information perception, information processing, and input behavior.
Figure 3 illustrates an example of how digital behaviors were extracted from 24/7 monitoring data. In the figure, the user engaged in learning activities for Web design. To compute cumulative text, screenshots were first converted into texts. The number of unique words appearing in the screenshots was considered cumulative text (341 unique words in this example). Then, screenshots were converted from color to grayscale to compute cumulative entropy. The entropy was computed by summing the proportion of pixels with each grayscale color weighted by its logarithm (the cumulative entropy of five screenshots in the example is 185). In the example, there were three Web browser tabs used, and the user browsed three distinct Web pages in this example. Two Websites were opened. The user opened the Google search interface and entered a query, for which the average typing speed was 242 milliseconds per character.
Fig. 3.
Fig. 3. An example of digital behavior extraction. In the figure, the user is engaged in learning activities for web design. The left side illustrates screen monitoring and digital activity monitoring data. The right side shows examples of how behavioral measures were determined and extracted from the monitoring data. This is an illustrative example of data as, for privacy reasons, it is not possible to reveal actual data from the experiments.

3.6 Cognitive Measures

Table 5 presents the set of measures used to quantify cognitive ability. We focused on these cognitive abilities because they have been shown to influence user behavior in prior research [10, 13, 25, 33, 50]. All participants performed a battery of standardized tasks that measured their cognitive abilities, including psychomotor speed, processing speed, selective attention, and working memory. Pavlovia platform1 and PsychoPy framework2 were used for the implementation of the cognitive testing tasks (see Appendix A.2 for details of these tasks). The cognitive measures and tasks are described as follows:
Psychomotor speed was measured by the digit symbol substitution test [21] in which the coordination of visual perception and finger movement was the core component of this task. A participant was instructed to fill as many empty boxes as possible with a symbol matching each digit. The score was the number of correct digit-symbol matches achieved in 120 seconds. The maximum score (or the total number of boxes) was 144.
Processing speed was measured by the pattern comparison tasks [6]. Participants had to judge if a pair of two stimuli, such as two letter strings or two patterns, were the same or not as fast as they could. A participant was given 90 seconds to respond to as many pairs as possible. The total number of pairs was 130. The score was the number of pairs that were identified correctly by the participant.
Selective attention was measured by stroop color word association test [24], which assessed cognitive flexibility and attention span by examining a participant’s ability to separate word and color naming stimuli. A participant was instructed to identify and name the colors of the presented words (ignoring the meaning of the words). A participant was given 45 seconds to respond to as many words as possible (up to a maximum of 100 words). The score was computed as the number of words that were named with the correct ink color.
Working memory was measured by an automated version of the operation span task [24, 59]. Participants needed to memorize and recall a set of letters for each trial while alternatively solving some simple math questions. The score is based on the operation span absolute score computation [59]. If a participant perfectly recalled all the letters in a single trial, his/her score was the number of letters (otherwise, 0 scores) for that trial. The score was computed by adding up the number of letters in all perfectly recalled trials. The score was normalized by dividing by the total number of letters in all the trials. For example, if there were three trials: trial 1 had a set of three letters, trial 2 had a set of four letters, and trial 3 had a set of five letters, and a participant correctly recalled the whole set of three letters in trial 1, the whole set of four letters in trial 2, and only three out of five letters in trial 3. The score was computed as 3 + 4 + 0 = 7 (zero score for trial 3 because not all letters were recalled correctly).
Fluid intelligence was measured using trailing making tasks [4, 33]. Participants were shown 25 circles on the computer screen containing numbers (1–13) and alphabets (A–L). The task was to find the rule that connects circles. To complete the task, the participant had to click on the circles in ascending patterns alternating between numbers and alphabets (1-A-2-B-3-C and so on) starting from the number 1. We measured the time taken to finish the test correctly.
Table 5.
Cognitive abilitiesDefinitionsTest measuresReferences
Psychomotor speedThe ability to actively maintain and manipulate information over a brief period of time.Digit symbolDigit symbol substitution test [21]
Processing speedThe ability to perceive, understand, and respond to the information.Pattern comparison scoreThe NIH toolbox pattern comparison processing speed test [6]
Selective attentionThe ability to focus on a particular object for a brief period, simultaneously ignoring distractions and irrelevant information.Stroop test— color wordStroop color word association test [24]
Working memoryA limited capacity store for retaining information for a brief period while performing mental operations on that information.Working memory capacityOperation span task [24]
Fluid intelligenceReasoning ability and the ability to generate, transform, and manipulate different types of novel information in real-time.Trail making performanceTrail making test [48]
Table 5. Cognitive Abilities and the Corresponding Test Measures
To avoid possible confounds in relation to fatigue, the cognitive testing session was conducted in the morning on a different day from the digital activity monitoring with a 5 minute break following each test. In addition, after the third test, there was a longer break in which participants were allowed to take as much time as they needed. There were instructions and practice trials prior to each test (except the intelligence test). Participants did not know which measures the tests involved and were asked to do the tests as best as they could following the instructions. Participants performed the tests in a randomized order, approximately 5–7 minutes per test. A cognitive testing session was conducted after the monitoring so that participants were unaware that their digital behavior would be associated with their cognitive abilities. Participants received no feedback or knowledge of results on these tests throughout the cognitive testing session but were debriefed upon exiting the study.

4 Associating Digital Behavior with Cognitive Abilities

Analyses were conducted to investigate the associative and causal relationships between cognitive abilities and digital behaviors. Here, we describe the procedure for the analyses, results, and discuss relationships.

4.1 Analysis Procedure

Regression analyses and path analyses were conducted to reveal associations and causal dependencies among behavioral factors and cognitive abilities. Shapiro–Wilk tests were conducted to determine whether or not the behavioral data followed a normal distribution.
Linear regression was used with measures of behavioral factors as dependent variables and measures of cognitive abilities as independent variables. All \(p\) values were adjusted using Bonferroni correction.
Path analysis was conducted to test causal relationships for the three behavioral factors using structural equation modeling [57]. In the model, the five cognitive measures were set as the independent variables. Measures of the behavioral factors were set as the dependent variables. Then behavioral factors: information perception, information processing, and input behavior were added as mediating variables in the model.

4.2 Results

The summary statistics of cognitive measures and behavioral measures are shown in Table 6. The results show that cognitive ability varies across individuals and large variances are observed for most cognitive measures indicating a recording of data from a diverse pool of participants.
Table 6.
FactorsMeasuresMinMaxMeanSD
Cognitive abilitiesPsychomotor speed21955320
Processing speed511148117
Selective attention15744014
Working memory19655210
Fluid intelligence4430611878
Digital behaviorsCumulative text per hour (words)8474,1552,337665
Cumulative entropy per hour185920508224
Tab change per hour11542911
Page visit per hour327146
Site visit per hour2942
 Omnibox typing speed (milliseconds)101591246125
Table 6. Summary Statistics for Measures of Cognitive Abilities and Behavioral Factors
Information Perception. Table 7 and Figure 4 show the results for the associations between cognitive abilities and the users’ information perception behavior. We found substantial effects of psychomotor speed and fluid intelligence on the amount of cumulative text consumed by the users. Regression analyses reveal an association between psychomotor speed and cumulative text rate with a coefficient \(F(1,18)=19.42\) and an effect size with \(R^{2}=0.533\) (\(p\,{\lt}\,0.001\)). Similarly, the results show a significant linear relationship between fluid intelligence and cumulative text per hour with an effect size (\(F(1,18)=13.21\), \(R^{2}=0.43\), \(p=0.002\)). The Shapiro-Wilk test indicated that cumulative text followed a normal distribution (\(W=0.95,p=0.39\)).
Table 7.
Digital behaviorsCognitive abilities\(\boldsymbol{F(1,18)}\)\(\boldsymbol{R^{2}}\)\(\boldsymbol{p}\)
Cumulative text per hourPsychomotor speed\(\boldsymbol{19.42}\)\(\boldsymbol{0.533}\)\(\boldsymbol{\,{\lt}\,0.001}\)
Processing speed\(2.63\)\(0.13\)\(0.12\)
Selective attention\(1.51\)\(0.082\)\(0.23\)
Working memory\(1.92\)\(0.1\)\(0.18\)
Fluid intelligence\(\boldsymbol{13.21}\)\(\boldsymbol{0.43}\)\(\boldsymbol{0.002}\)
Cumulative entropy per hourPsychomotor speed\(1.64\)\(0.088\)\(0.21\)
Processing speed\(0.37\)\(0.021\)\(0.55\)
Selective attention\(0.15\)\(0.0092\)\(0.7\)
Working memory\(0.21\)\(0.12\)\(0.66\)
Fluid intelligence\(1.8\)\(0.09\)\(0.19\)
Table 7. Coefficient Values and Effect Sizes Found from Linear Regression Analyses for Information Perception
Bold-faced values indicate statistically significant effects (\(p\,{\lt}\,0.05\)). All p values were adjusted for multiple comparisons.
Fig. 4.
Fig. 4. Linear regression relationship between information perception and five cognitive abilities.
Figure 4(a) also indicates an association between the amount of cumulative text consumed per hour and the measures of each of the two cognitive abilities (psychomotor speed and fluid intelligence). Generally, regression tests revealed that individuals who scored higher in psychomotor speed (higher in the speed of thinking and decision-making) and higher in fluid intelligence were able to scan texts on the screen more quickly than those who scored lower.
Furthermore, we examined the effects of cognitive abilities on the amount of cumulative entropy, as shown in Figure 4(b). Significant associations between the amount of cumulative entropy per hour and five cognitive abilities were not observed.
Information Processing. Table 8 and Figure 5 show the associations between cognitive abilities and the user’s information processing behavior. For tab-changing behavior, three cognitive abilities: processing speed, selective attention, and fluid intelligence, were associated with the tab-change rate with substantial effect sizes. Regression analysis revealed an association between processing speed and tab change rate with a coefficient \(F(1,18)=8.604\) and an effect size \(R^{2}=0.34\) (\(p=0.0093\)). Similarly, \(F(1,18)=8.61\) and \(R^{2}=0.34\) (\(p=0.0092\)) were found for selective attention ability; and \(F(1,18)=6.28\) and \(R^{2}=0.35\) (\(p=0.023\)) were found for fluid intelligence. The Shapiro-Wilk test revealed that the tab changing rate followed a normal distribution (\(W=0.91,p=0.25\)).
Table 8.
Digital behaviorsCognitive abilities\(\boldsymbol{F(1,18)}\)\(\boldsymbol{R^{2}}\)\(\boldsymbol{p}\)
Tab change per hourPsychomotor speed\(3.21\)\(0.16\)\(0.091\)
Processing speed\(\boldsymbol{8.604}\)\(\boldsymbol{0.34}\)\(\boldsymbol{0.0093}\)
Selective attention\(\boldsymbol{8.61}\)\(\boldsymbol{0.34}\)\(\boldsymbol{0.0092}\)
Working memory\(2.17\)\(0.11\)\(0.16\)
Fluid intelligence\(\boldsymbol{6.28}\)\(\boldsymbol{0.35}\)\(\boldsymbol{0.023}\)
Page visit per hourPsychomotor speed\(2.38\)\(0.12\)\(0.14\)
Processing speed\(\boldsymbol{5.98}\)\(\boldsymbol{0.26}\)\(\boldsymbol{0.026}\)
Selective attention\(3.16\)\(0.16\)\(0.093\)
Working memory\(1.96\)\(0.1\)\(0.18\)
Fluid intelligence\(3.81\)\(0.18\)\(0.06\)
Site visit per hourPsychomotor speed\(0.52\)\(0.03\)\(0.48\)
Processing speed\(0.29\)\(0.017\)\(0.59\)
Selective attention\(\boldsymbol{5.18}\)\(\boldsymbol{0.23}\)\(\boldsymbol{0.036}\)
Working memory\(0.001\)\(0.001\)\(0.99\)
Fluid intelligence\(0.04\)\(0.01\)\(0.83\)
Table 8. Coefficient Values and Effect Sizes of the Linear Regression Analyses for Information Processing
Bold-faced values indicate statistically significant effects (\(p\,{\lt}\,0.05\)). All p values were adjusted for multiple comparisons.
Fig. 5.
Fig. 5. Linear regression relationship between information processing and five cognitive abilities.
Figure 5(a) presents an illustration of the regression analysis for the tab change rate against measures of the five cognitive abilities. Linear relationships between the number of tabs that changed over time and the three cognitive abilities were found: processing speed, selective attention, and fluid intelligence. Intuitively, this result indicates that users with lower processing speed, lower attention ability, and lower fluid intelligence spent more time on individual tabs than their counterparts who obtained higher scores on these cognitive tests.
Figure 5(b) shows the effects of five cognitive abilities on the user’s page visit behavior. Associations were found between processing speed and page visit rate (\(F(1,18)=5.98\), \(R^{2}=0.26\), \(p=0.026\)) and found to be normally distributed with \(W=0.97,p=0.87\) for page visit rate. The results indicate that users with higher processing speed visited Web pages more often than users with lower processing speed. In contrast, users with lower processing speed tended to spend more time dwelling on the pages they accessed.
Figure 5(c) shows the effects of five cognitive abilities on the site visit behavior. Interestingly, we found that only selective attention ability was associated with this behavior with a coefficient \(F(1,18)=5.18\) and an effect size \(R^{2}=0.23\) (\(p=0.036\)). The Shapiro-Wilk test also revealed that site visit behavioral measure was normally distributed with \(W=0.97,p=0.81\). The results suggest that individuals with better selective attention could focus on relevant information better while moving between Websites.
Input Behavior. Table 9 and Figure 6 present the results of regression analyses between cognitive abilities and Omnibox typing speed. We found an effect of working memory on how fast the users type in the Omnibox (address bar and search box). The effect was substantial with a coefficient \(F(1,18)=5.033\) and \(R^{2}=0.23\) (\(p=0.038\)). The Shapiro-Wilk test indicated that input behavior data was normally distributed with \(W=0.91\) and \(p=0.21\). The results suggest that people who scored lower in the working memory test would take longer to type in the Omnibox. The reason is likely due to the need to recall a Website URL or the content of a query, which may depend on a user’s working memory.
Table 9.
Digital behaviorsCognitive abilities\(\boldsymbol{F(1,18)}\)\(\boldsymbol{R^{2}}\)\(\boldsymbol{p}\)
Omnibox typing speedPsychomotor speed\(0.38\)\(0.02\)\(0.54\)
Processing speed\(1.53\)\(0.083\)\(0.23\)
Selective attention\(0.037\)\(0.0022\)\(0.85\)
Working memory\(\boldsymbol{5.033}\)\(\boldsymbol{0.23}\)\(\boldsymbol{0.038}\)
Fluid intelligence\(0.75\)\(0.04\)\(0.39\)
Table 9. Coefficient Values and Effect Sizes Found from Linear Regression Analyses for Input Behavior
Bold-faced values indicate statistically significant effects (\(p\,{\lt}\,0.05\)). All p values were adjusted for multiple comparisons.
Fig. 6.
Fig. 6. Linear regression relationship between Omnibox typing speed and five cognitive abilities.
Causal Effects. The linear regression analyses revealed how cognitive differences were associated with multiple behavioral factors. We considered structural equation modeling for path analysis to further understand the causal dependencies of digital behaviors on cognitive abilities. Because prior works have shown possible relations between cognitive abilities [12] that may explain the direct and indirect effect on user behavior, we also report results using a path model that accounts for correlation and causality between the abilities. Additionally, the path model also revealed associations among behavioral factors, such as information processing, information perception, and input behavior.
Figure 7 presents the results of structural equation models with each node showing a coefficient for each link. Nodes with links in between indicate significant dependencies with \(p\,{\lt}\,0.05\), and the arrows indicate causal directions. In contrast, nodes without links in between indicate non-significant relations.
Fig. 7.
Fig. 7. Path models describe causal relations between cognitive abilities and behavioral factors. Independent variables (cognitive abilities) are on the left, dependent variables (behavioral factors) are on the right, and those are illustrated as squares. Mediating variables (information processing, information perception, and input behavior) are illustrated as circles. Nodes with links in between indicate significant relations, and nodes without links in between indicate non-significant relations.
The structural equation models confirmed the results of regression analyses. Selective attention was associated with the amount of information processed by the users and also caused individual differences in this behavior—-with a coefficient of 7.73 (\(p=0.009\)). This suggests that higher selective attention causes a better ability to process information from multiple sources, including increased tab change rate and the number of page and site visits per hour. For example, for users with higher selective attention abilities, the cost of switching between tabs and Web pages (the ability to maintain focus on important information while changing tabs) is much lower than for those with lower attention spans. We also found that processing speed and fluid cognitive abilities indirectly affected how users processed information.
Fluid intelligence has a coefficient of 0.17 (\(p=0.03\)), meaning that higher intelligence causes moderately better perception. Users with higher fluid intelligence were faster at scanning textual and visual information than those with lower fluid intelligence. On the other hand, users with lower fluid intelligence could compensate for their lower cognitive ability with a more focused approach. For instance, they could slowly examine the on-screen content so as not to miss any important information, and carefully decide to change the tab so that each interaction would lead to useful information to maximize the information gained per tab unit. The path models confirmed a relationship between information processing behavior and information perception behavior with a coefficient of 0.44 (\(p=0.024\)). In addition, psychomotor speed has a coefficient of 0.51 (\(p=0.004\)) with fluid intelligence, indicating that psychomotor speed may have an indirect effect on the user’s information perception ability.
The results also showed a direct path between working memory and typing speed confirming that faster typing speed was caused by better working memory of individuals. A negative coefficient of \(-0.29\) (\(p=0.035\)) was found for the causal dependence of input behavior on working memory. Consequently, individuals with lower working memory capacity would type more slowly than those with larger working memory capacity. This result must be interpreted in context. The typing speed was measured from typing in the address bar and search boxes, which have a particular requirement of working memory access. Therefore, the result may not generalize to typing or writing more generally.
Furthermore, the path model revealed an association between information perception and information processing with a coefficient of 0.44 (\(p=0.024\)). This result indicates that the amount of information that a user processed was influenced by the number of pages the user opened. However, there were no associations among input behavior information perception, and information processing.

5 Predicting Cognitive Abilities

One of the main goals of our research was to predict the cognitive abilities given the user’s digital behavioral data. We describe the prediction setup and results below.

5.1 Prediction Setup

We considered both linear and non-linear approaches for the prediction. Two prediction models were chosen: a Random Forest model and a Multiple Linear Regression model. These models have also been previously shown to be effective in similar tasks [13, 25]. Behavioral variables were used to construct a model for predicting cognitive ability.
The models were trained with leave-one-user-out cross-validation, and the data were split into a training set and a test set accordingly. In each iteration, the model was trained using data from 19 participants and predicted the cognitive measures of the left-out participant, such that the test set always contains an unseen user. More specifically, the input for training the models consisted of feature vectors of digital behaviors (cumulative text, cumulative entropy, tab change rate, page visit, site visit, and Omnibox typing speed) and the corresponding cognitive test scores of 19 participants. The model output was the prediction of individual cognitive test scores of the left-out participant—as the experiment followed a cross-validation setup. The performance of the models was then evaluated using Root Mean Squared Error (RMSE) for all cognitive measures. RMSE is a commonly used measure in the field of machine learning and data science [1, 20]. The RMSE was calculated by taking the difference between the predicted values and the actual values. The results were then averaged and the square root of the average was taken. This metric indicates how close the predicted values were to the actual values. Lower values indicate better performance.
For Random Forest models, the hyper-parameter max tree depth and the number of trees were determined by 10-fold cross-validation using the training set within each test user separately. To set the hyper-parameter of Multiple Linear Regression models, we enabled fit intercept and disabled normalization. This ensured that data leakage could be fully avoided as the model had no information about the user’s data used for evaluation.
The models were compared with two control models: (1) prediction using a permutation technique and (2) random guessing/prediction. For the permutation technique, we randomly permuted all features (behavior factors) and targets (cognitive measures) on the training set so that any possible relation between features and targets, except their distribution, was destroyed when constructing the model. Thus, this training procedure retains the distribution of the data but should not carry any predictive information. We used these permuted data to fit the Random Forest model for prediction. For each LOO iteration, the permutation process was repeated 1,000 times to get a vector of 1,000 values of RMSE, and the mean of this vector was used as the output value. For random guessing/prediction, we generated a random value for a measure of cognitive ability. We also repeated the random guessing process 1,000 times. We used paired t-tests with Bonferroni correction to test for significant differences between the models.

5.2 Results

Figure 8 shows the results for cognitive ability prediction. The Random Forest model that leveraged all available features produced the lowest RMSE in all cognitive abilities. Pairwise comparisons indicate significant differences in RMSE between the Random Forest model and the control models (permutation and random guessing/prediction), with all \(p\) values \(\,{\lt}\,\)0.007.
The Multiple Linear Regression model also produced small RMSEs compared to the control models. Significant differences were found between the Multiple Linear Regression model and the two control models (permutation and random guessing/prediction) for psychomotor speed and selective attention, with \(p\) values \(\,{\lt}\,\)0.007. However, using Multiple Linear Regression to predict processing speed and working memory did not produce an RMSE better than the permutation models.
Fig. 8.
Fig. 8. Results of cognitive ability prediction in terms of RMSE. RFr, Random Forest; MLR, Multiple Linear Regression; Per, Permutation; and RPr, Random Prediction. All p values were adjusted for multiple comparisons. All p values were adjusted for multiple comparisons.
The permutation models also show relatively low error rates because the cognitive ability measures seem to have a natural range. In other words, cognitive measures do not vary between 0 and a maximum score but are on a more limited range reflecting empirical lower-bound and upper-bound cognitive abilities of healthy users. However, all results show significant differences between the permutation and prediction models trained with non-permutated data. In summary, the results suggest that cognitive abilities can be predicted reliably based on behavioral patterns with low error rates, and significantly outperforming permutation and random models.
Figure 9 presents the importance of behavioral features and their contributions in the Random Forest models. The results indicate that cumulative text was the most important feature for predicting psychomotor speed, processing speed, and fluid intelligence. Information processing behavior comprising tab change rate, page visit, and site visit were the most important features for predicting selective attention. Lastly, typing speed was among the top features for predicting working memory. The results of feature importance are consistent with regression analysis and path model analysis that working memory, selective attention, and fluid intelligence affect users’ digital behavior.
Fig. 9.
Fig. 9. Feature importance: contribution weights in Random Forest. Feature importance is calculated as the decrease in node impurity weighted by the probability of reaching that node. The node probability is calculated by the number of samples that reach the node divided by the total number of samples. The higher the value, the more important the feature.

6 Discussion

6.1 Summary of Findings

The results show how cognitive abilities can explain differences in digital behaviors and how the use of behavioral patterns predicts cognitive abilities. In this section, we distill the findings and answer the RQs defined earlier.
RQ1. How Is Cognitive Ability Associated with Naturalistic Digital Behaviors? The results of regression analyses and structural equation modeling indicate that different measures of cognitive ability are associated with features of individuals’ digital behavior. We found that higher fluid intelligence was associated with and caused faster information processing. More specifically, people having higher fluid intelligence processed information, such as scanned text, faster than those who scored lower. In addition, our results suggested an indirect effect of psycho-motor speed on fluid intelligence. Higher selective attention was associated and had a causal relationship with the amount of information processed by individuals. Participants with higher selective attention changed tabs more frequently. This finding is congruent with findings from previous research that linked cognitive abilities, and multitasking [44]. On the other hand, individuals who scored lower in selective attention spent more time examining information on individual tabs. Findings in recent work [13] revealed that lower perceptual speed negatively affected the ability to perceive visual information. Thus, our results suggest an extension to the previous findings by determining that selective attention could also affect the user’s information perception. We also found that higher working memory capacity was associated and had a causal relationship with typing speed. Participants with higher working memory capacity typed faster than participants with lower working memory capacity.
RQ2. Can We Predict Cognitive Abilities from Naturalistic Digital Behaviors? We found that predictive models utilizing all behavioral features significantly outperformed control models for all five cognitive abilities. The Random Forest model yielded the best performance with the smallest error rates compared to the other models. The results indicate that behavioral and interaction data are predictive of users’ cognitive abilities and applicable to naturalistic data obtained in the wild. This indicates that naturalistic behavioral data can be utilized to estimate users’ cognitive ability. We also found that using a combination of behavioral features can be beneficial for predicting cognitive abilities. Different cognitive abilities can be predicted using a different set of behavioral cues. For example, working memory can be predicted using a combination of typing speed and tab change rate. Similarly, fluid intelligence can be predicted using a combination of the frequency of page visits and the amount of perceived cumulative text.

6.2 Implications

We found that the cognitive abilities of individuals have a direct association with their daily digital activities. While a recent work shows that differences in computer usage may be primarily attributable to differences in user preference [22], our analyses revealed that many behavioral differences could be largely caused by cognitive ability. Our findings suggested that digital behavior could be used as a passive preliminary detection tool for cognitive ability and even impairment.
The results of our prediction strongly indicated the feasibility of modeling and predicting cognitive abilities. In particular, a simple model, such as Random Forest yielded low error rates for predicting cognitive abilities. Currently, most cognitive assessment is done in clinical trials when a person visits a physician, but our study revealed it may be possible to detect cognitive decline directly from digital behavior sources, opening new opportunities for researchers in cognitive sciences and possibly even clinical practice.
A generalizable, predictive model can also save time and cost for user interface designers or researchers by offering them prior insights into individuals’ backgrounds and interaction difficulties before testing with the users. Although our results provide initial guidance on how to build compelling user models depending on the target cognitive abilities, there is a need to examine whether these findings would apply to other modeling tasks beyond the ones analyzed here.
Detection of cognitive abilities may also benefit users when coupled with an adaptive system [69]. Various designs and techniques may be chosen to adapt to individual users’ cognitive abilities. In particular, adaptive systems and services can benefit people affected by the high intensity of information presentation and the requirement of memory recall for input. For instance, users with reduced ability to maintain focus on relevant pieces of information (low selective attention) could benefit from customized user interface layouts that prioritize essential information while removing other distractions or extraneous details. For users who have low fluid intelligence, because those people have difficulty figuring out rules operating user interfaces; less dense text designs featuring more visual metaphors and navigation aids could be implemented. For users with reduced working memory who had slow typing speed, interfaces could be designed with simplified navigation-based menu options in which the information could be accessed through shortcuts rather than manually typing it in. This approach could lower the cognitive load in retrieving the information.
The error rates of the predictive models were low, and we believe the prediction accuracy is sufficient for transferring these results to practical systems. Nonetheless, given that there are still small errors in predicting the user’s cognitive ability, researchers should consider that the benefit of supporting users who need help might still outweigh the possible hindrance caused by unwarranted adaptation delivered to a minority of the users. However, further experiments should investigate the findings in the scope of real-world adaptive systems to enable useful interventions and reduce the possible intrusiveness of unwarranted adaptation. For instance, personalization could be suggested to the users via a gentle prompt message, in a mixed-initiative fashion, so that users could still ignore or decline the adaptations if they do not perceive them as useful.

6.3 Limitations

Sample Size. The sample of participants was relatively small and consisted primarily of knowledge workers. Nevertheless, it was carefully designed with the assumption that knowledge workers would use computers more often than other groups and that the varying ages of participants would likely account for cognitive differences among them. However, results were found to be significant even after correcting for multiple comparisons in a very conservative way (Bonferroni). Moreover, the predictive models also performed better than random prediction, which indicated that data was sufficient to reveal the suggested associations and to demonstrate that prediction was possible. In addition, our sample is not fully representative of the computer-using population, and our results should be interpreted in this context. Future work could consider a large-scale data collection campaign spanning different user types and target populations to provide further evidence on computer usage patterns.
Behavioral Recordings Source. We focused on behavioral signals recorded on a personal computer and did not consider data from other devices, such as smartphones. More and more diverse data could complement the present findings and possibly improve the prediction results. Similar to interaction with mobile devices [25], digital behavior in computer systems also depends on many contextual factors, which could become additional confounding factors for data analysis. For instance, the data collected did not account for individual device quality. Because our experiments were conducted in the wild, device factors, such as screen size, speed, and ergonomics could potentially impact how participants used their computers. Nevertheless, we have already taken an active step toward this limitation by involving participants in completing a prescreening questionnaire, reporting that they have been using computers for daily work and leisure activities and were familiar and comfortable with their devices. Therefore, it is unlikely that these factors would have affected the significance of our findings.
In addition, digital behavior captured for this research occurred within the naturally occurring context of their work and home activities. It is difficult to ascertain the full scope of influences on digital behavior in such an environment. Despite this factor, we believe that data captured from the real-world environment contributed to the strength of our findings. Our studies can better explore how individuals understand, use, and appropriate technologies in real-world situations. The possibility of linking such behavior with cognitive ability could open up a new direction for understanding the cognitive aspects of HCL. For instance, predicting cognitive ability could benefit personalized adaptive systems, such as designing cognition-aware technologies focusing on ways to minimize cognitive effort for interface interaction by reducing the complexity of user interfaces, such as adjusting the amount of information presented to users having shorter span of attention or recommending browsing histories [29, 64, 65] to aid users having reduced working memory.
Behavioral Factors. Our analyses only concerned a limited number of behavioral factors since they were used in previous research, e.g., [19, 22, 34, 63, 66, 72]. We accounted for different tasks and habits by normalizing the measures per hour so that measures would be comparable and independent of temporal and task variability. However, future work could consider other metrics of digital behaviors, such as re-visitation patterns, sequential patterns of application use, or contextual factors from other devices (e.g., mobile). Furthermore, behavioral metrics used in prior works were mostly Web browser-specific [8, 9, 13, 17]. In this study, we also considered other behavioral factors that were independent of specific application sources, including cumulative text and entropy. Future research could investigate how the models could benefit from non-browser behavioral data to predict the cognitive abilities of users who do not use the Web frequently.
In addition, only Shannon theory [49] was utilized for the calculation of cumulative entropy; future work could consider color entropy and luminance entropy of an image [72]. Cumulative text measure only accounts for new text appearing on the screen. However, participants might not necessarily focus on all the words in the text, or participants would just change tabs without reading the content. However, prior research has considered these behavior signals as useful indicators of differences in information processing and perception across different demographic populations [68, 72]. They are easy to obtain because current logging systems in many modern web applications are able to capture these types of signals, while sophisticated approaches, such as web-based eye tracking are costly, require calibration, and sometimes encounter certain problems with external variables, such as varying lighting and device configurations, as well as privacy issues. Our research focused on accessible metrics, such as the amount of text presented on the screen and the findings show that these indicators are sufficient for revealing meaningful relationships. We are also aware that these measures do not perfectly capture the user’s information processing activity. To address the limitations of the measures, in addition to cumulative text and entropy, we also considered several moderating factors, such as user engagement and interaction. For instance, we have used a variety of measures involved with reading, such as time spent on Web pages and typing behavior, to better understand the degree to which users process and use information from Web pages. Our research has shown that the combination of these measures and moderating variables offers considerable insight into how people process and perceive Web pages connecting to their cognitive ability. However, further research is required to better understand the accuracy of our measurements compared to the accuracy of alternative measures, such as eye tracking, scrolls, and clicks.
Some measures assigned to behavioral factors may not entirely represent the characteristics of those factors but fractions of them. For instance, the Omnibox input may be limited but is still indicative of more general input behavior from the user. Furthermore, modern browsers often autocomplete the URLs and limit the number of characters that the user has to input. This limits the extent of input behavior that Omnibox inputs can represent. However, it was also the aim of this research to investigate how such commonly used measures could reveal cognitive ability. In addition, those behavioral signals might be affected by the user’s physical environment, such as sitting position and ambient noise that could affect the speed of typing queries or URLs. It might take days or weeks of monitoring for the model to learn from this behavior signal to successfully predict cognitive abilities, while it would take less data (e.g., an hour or so of monitoring) when a user works on the computer in a lab-based controlled environment, where the user is provided with a proper chair and a quiet condition. This is also an important implication as we demonstrated that a larger portion of data was necessary to obtain significant results. However, the duration of the experiment, of course, was determined in the design phase and it may be possible that shorter monitoring would also be enough. Future studies could examine how the predictive accuracy of cognitive abilities would be affected with respect to the length of the monitoring phase. This is to improve user privacy and reduce the amount of data collected.
Data Collection. Users could turn off the data collection software anytime during the study. This feature was necessary for ethical data collection as participants consented with the option to opt out of the study at any time. It also allowed for a potential bias in the collected data, as participants could disable data collection on sensitive Websites. However, we observed that the screen monitoring was operational well over 95% of the time and was turned off on average less than 30 minutes per participant. Consequently, we believe that the data used for the analysis represents naturalistic digital behavior. Furthermore, because age has been found to be associated with cognitive ability, future studies could also explore the potential for predicting age from digital behavior.

6.4 Ethical Considerations

The ability to detect cognitive abilities from simple computer usage offers a powerful tool for understanding users and adapting computation to make information more accessible for users with varying cognitive abilities. At the same time, the finding that simple behavioral data can predict cognitive ability raises severe privacy concerns.
Our results indicate that as relatively simple behavioral measures could allow us to predict cognitive abilities, we should be careful about the kind of data we allow service providers to collect via Web browsers or other applications. Cognitive profiles can be used to derive the user’s identity and cognitive ability with little or no consent from users. This means that data-sharing policies should be put in place that restrict the use of user behavioral data for predicting cognitive ability. Users should be given control over the data they are willing to give out for analysis, such as control of what goes into the user model, what information from their model is available to different services, and how the model is managed and maintained.
We followed strict guidelines and policies also in the research reported here. As monitoring data can reveal the user’s cognitive ability to unauthorized third-party entities, we were well aware of such implications and carefully managed users’ data with encryption, controlling access, and building models so that private data was not available even for the researchers themselves. Our study procedure and consent forms were reviewed and ethically accepted before data collection. All analyses were conducted offline, and data was not exposed to undesired third parties.

7 Conclusion

We set out to study whether cognitive abilities are associated with and predictable from digital behavior. Findings in early works on predicting user cognitive abilities focused on the use of eye-tracking, while our research examined other measures, such as the amount of text viewed on the screen or time spent on tabs and pages. Our findings showed that these non-intrusive signals were able to reveal meaningful associations with user cognitive abilities and could potentially improve prediction performance when combined with eye-tracking. In addition, most prior works focused on prediction tasks for a limited number of cognitive abilities. Here, we showed the feasibility of predicting measures for varying cognitive abilities from naturalistic digital behaviors captured in everyday digital activities. Associative, path models, and predictive models were utilized to reveal insights into the relationships and predictability of digital behavior traits for a wide set of variables representing cognitive ability. All in all, our research demonstrates the possibility of predicting cognitive abilities from everyday digital activity. The prediction may eventually help reduce the efforts researchers have to put in for estimating cognitive ability and would benefit future research for better user modeling, personalization, and cognitively adaptive computing.

Authors’ Statement

This contribution is original and has not been previously published. All the text, user study, and analysis contained in this article are unpublished and there are no concurrent submissions.

Footnotes

Appendices

A.1 Pre-Screen Questionnaire

The questionnaire was developed by the research team and attached to the recruiting message (Table 10). Only respondents who had sufficient skills and experience in using computers were considered eligible for the study.
Table 10.
Demographics and health—Age (years)
 —Gender (male/female)
 —Education: high school, vocational education, bachelor’s, master’s degree
 and above (multiple choices)
 —Do you have a normal or corrected-to-normal vision? (yes/no)
 —Do you have—or have you had—a diagnosed, on-going mental illness?
 (yes/no)
 —Are you currently taking any medication to treat symptoms of depression
 or anxiety? (yes/no)
Computer specification—Model name of your laptop? (free-text)
 —What OS is on your laptop? (free-text)
 —Are you comfortable with using your laptop? (yes/no)
Computer use and experience—Years of computer use (e.g., computer experience in general)? (years)
 —Purpose of laptop use: home, work, or home/work? (multiple choices)
 —Hours of use per day? (hours)
Computer skill—What applications used for composing textual documents? (free-text)
—What applications are used for preparing a slideshow presentation? (free-text)
—What applications are used for sending/receiving emails? (free-text)
—What applications used for instant messaging (free-text)
—Describe what a Web browser is? (free-text)
—Describe what a search engine is? (free-text)
—Name of a Web browser application do you usually use? (free-text)
—Name of a search engine on the laptop do you usually use? (free-text)
—What does the order of search results on a search interface mean? (free-text)
Computer task—Following a guide for installing a software on your laptop.
Table 10. Questionnaire

A.2 Cognitive Tests

Psychomotor Speed was measured using Digital Symbol Substitution Test. Figure 11 demonstrates the task. The participant sees on a computer screen a grid of nine symbol-number matching pairs on the top of the screen. Given a symbol, the participant is instructed to push the right answer key (number).
Fig. 10.
Fig. 10. Digital symbol substitution test.
Fig. 11.
Fig. 11. Examples of pattern comparison processing speed test.
Processing Speed was measured using Pattern Comparison Test. Figure 11 demonstrates the task. This test required participants to identify whether two visual patterns are the “same” or “not the same” (responses were made by pressing a “yes” or “no” button).
Selective Attention was measured using stroop test. Figure 12 demonstrates the tasks. In the stroop Test, participants simply look at color words, such as the words “blue,” “red,” or “green,” the task is to name the color of the ink the words are printed in, while fully ignoring the actual word meaning.
Fig. 12.
Fig. 12. Examples of stroop test.
Fig. 13.
Fig. 13. Examples of operation span task.
Working Memory was measured using operation span task. Figure 12 demonstrates the tasks. Participants try to remember sequentially presented words in their correct order while simultaneously solving simple math equations.
Fluid Intelligence was measured using Trail Making Tasks. Figure 14 demonstrates the tasks. Participants need to figure out the rule to connect circles (numbers and alphabets). The rule is 1-A-2-B-3-C…13-F.
Fig. 14.
Fig. 14. Examples of trail making task.

A.3 Application Categories

The application categories were based on the prior work [35]. Applications were manually coded by the research team (Table 11).
Table 11.
CategoryTypes of applications
Communication—Email, chat clients, video conferencing
File system and storage—File explorer, cloud storage, file transfer
System operations—Installing and uninstalling programs
—Download managers
—Notification windows
—Start menu
—Antivirus, malware removal tools
— Display, audio, power, and network settings
—Data backup, data recovery, system recovery
Web browsing—Chrome, firefox
Office and work—Billing, payroll, accounting tools
—Program editors, IDEs, debuggers
—Text, word, spreadsheet, and presentation editors
—Productivity utilities, data analysis software
—3D modeling, scientific imaging and sensors
Music and video—Streaming software
—Video players
Table 11. Applications That Participants Used on Their Computer

References

[1]
Robert St. Amant, Thomas E. Horton, and Frank E. Ritter. 2007. Model-Based Evaluation of Expert Cell Phone Menu Interaction. ACM Transactions on Computer-Human Interaction 14, 1 (May 2007), 1–es. DOI:
[2]
Anne Aula, Rehan M. Khan, and Zhiwei Guan. 2010. How Does Search Behavior Change as Search Becomes More Difficult? In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (CHI ’10). ACM, New York, NY, 35–44. DOI:
[3]
Shirley Ann Becker. 2004. A Study of Web Usability for Older Adults Seeking Online Health Resources. ACM Transactions on Computer-Human Interaction 11, 4 (Dec. 2004), 387–406. DOI:
[4]
Riccardo Brunetti, Claudia Del Gatto, and Franco Delogu. 2014. eCorsi: Implementation and Testing of the Corsi Block-Tapping Task for Digital Tablets. Frontiers in Psychology 5 (Aug. 2014), 939. DOI:
[5]
Hancheng Cao, Chia-Jung Lee, Shamsi Iqbal, Mary Czerwinski, Priscilla N. Y. Wong, Sean Rintel, Brent Hecht, Jaime Teevan, and Longqi Yang. 2021. Large Scale Analysis of Multitasking Behavior During Remote Meetings. In Proceedings of the CHI Conference on Human Factors in Computing Systems (CHI ’21). ACM, New York, NY, Article 448, 13 pages. DOI:
[6]
Noelle E. Carlozzi, Jennifer L. Beaumont, David S. Tulsky, and Richard C. Gershon. 2015. The NIH Toolbox Pattern Comparison Processing Speed Test: Normative Data. Archives of Clinical Neuropsychology 30, 5 (May 2015), 359–368. DOI:. Retrieved from https://academic.oup.com/acn/article-pdf/30/5/359/4979718/acv031.pdf
[7]
Xiuli Chen, Sandra Dorothee Starke, Chris Baber, and Andrew Howes. 2017. A Cognitive Model of How People Make Decisions Through Interaction with Visual Displays. In Proceedings of the CHI Conference on Human Factors in Computing Systems (CHI ’17). ACM, New York, NY, 1205–1216. DOI:
[8]
Jessie Chin and Wai-Tat Fu. 2010. Interactive Effects of Age and Interface Differences on Search Strategies and Performance. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (CHI ’10). ACM, New York, NY, 403–412. DOI:
[9]
Jessie Chin and Wai-Tat Fu. 2012. Age Differences in Exploratory Learning from a Health Information Website. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (CHI ’12). ACM, New York, NY, 3031–3040. DOI:
[10]
Jessie Chin, Wai-Tat Fu, and Thomas Kannampallil. 2009. Adaptive Information Search: Age-Dependent Interactions between Cognitive Profiles and Strategies. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (CHI ’09). ACM, New York, NY, 1683–1692. DOI:
[11]
Aleksandr Chuklin, Ilya Markov, and Maarten de Rijke. 2015. Click Models for Web Search. Synthesis Lectures on Information Concepts, Retrieval, and Services 7, 3 (2015), 1–115. DOI:
[12]
Aaron Cochrane, Vanessa Simmering, and C. Shawn Green. 2019. Fluid Intelligence is Related to Capacity in Memory As Well As Attention: Evidence from Middle Childhood and Adulthood. PloS one 14, 8 (Aug. 2019), 1–24. DOI:
[13]
Cristina Conati, Sébastien Lallé, Md Abed Rahman, and Dereck Toker. 2020. Comparing and Combining Interaction Data and Eye-Tracking Data for the Real-Time Prediction of User Cognitive Abilities in Visualization Tasks. ACM Transactions on Interactive Intelligent Systems 10, 2, Article 12 (May 2020), 41 pages. DOI:
[14]
Cristina Conati and Heather Maclaren. 2008. Exploring the Role of Individual Differences in Information Visualization. In Proceedings of the Working Conference on Advanced Visual Interfaces (AVI ’08). ACM, New York, NY, 199–206. DOI:
[15]
Rianne Conijn, Jens Roeser, and Menno Zaanen. 2019. Understanding the Keystroke Log: The Effect of Writing Task on Keystroke Features. Reading and Writing 32, 9 (Nov. 2019), 2353–2374. DOI:
[16]
Michael Crabb and Vicki L. Hanson. 2014. Age, Technology Usage, and Cognitive Characteristics in Relation to Perceived Disorientation and Reported Website Ease of Use. In Proceedings of the 16th International ACM SIGACCESS Conference on Computers & Accessibility (ASSETS ’14). ACM, New York, NY, 193–200. DOI:
[17]
Kyle Crichton, Nicolas Christin, and Lorrie Faith Cranor. 2021. How Do Home Computer Users Browse the Web? ACM Transactions on the Web (TWEB) 16, 1, Article 3 (Sept. 2021), 27 pages. DOI:
[18]
Sara Czaja, Neil Charness, Arthur Fisk, Christopher Hertzog, Sankaran Nair, Wendy Rogers, and Joseph Sharit. 2006. Factors Predicting the Use of Technology: Findings From the Center for Research and Education on Aging and Technology Enhancement (CREATE). Psychology and Aging 21 (Jul. 2006), 333–52. DOI:
[19]
Krista DeLeeuw and Richard Mayer. 2008. A Comparison of Three Measures of Cognitive Load: Evidence for Separable Measures of Intrinsic, Extraneous, and Germane Load. Journal of Educational Psychology 100 (Feb. 2008), 223–234. DOI:
[20]
Fernando Diaz and Rosie Jones. 2004. Using Temporal Profiles of Queries for Precision Prediction. In Proceedings of the 27th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR ’04). ACM, New York, NY, 18–24. DOI:
[21]
Lisa Whipple Drozdick, Susan Engi Raiford, Dustin Wahlstrom, and Lawrence G. Weiss. 2018. The Wechsler Adult Intelligence Scale—Fourth Edition and the Wechsler Memory Scale—Fourth Edition. In Contemporary Intellectual Assessment: Theories, Tests, and Issues. D. P. Flanagan and E. M. McDonough (Eds.), The Guilford Press, 486–511.
[22]
Patrick Dubroy and Ravin Balakrishnan. 2010. A Study of Tabbed Browsing among Mozilla Firefox Users. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (CHI ’10). ACM, New York, NY, USA, 673–682. DOI:
[23]
Geoffrey B. Duggan and Stephen J. Payne. 2011. Skim Reading by Satisficing: Evidence from Eye Tracking. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (CHI ’11). ACM, New York, NY, 1141–1150. DOI:
[24]
Charles J. Golden, Christie Golden, and Cs Golden. 1978. Stroop Color and Word Test: Manual for Clinical and Experimental Uses. Hakjisa.
[25]
Mitchell L. Gordon, Leon Gatys, Carlos Guestrin, Jeffrey P. Bigham, Andrew Trister, and Kayur Patel. 2019. App Usage Predicts Cognitive Ability in Older Adults. In Proceedings of the CHI Conference on Human Factors in Computing Systems (CHI ’19). ACM, New York, NY, 1–12. DOI:
[26]
James J. Heckman, Jora Stixrud, and Sergio Urzua. 2006. The Effects of Cognitive and Noncognitive Abilities on Labor Market Outcomes and Social Behavior. Journal of Labor Economics 24, 3 (2006), 411–482. DOI:
[27]
Juan Pablo Hourcade, Christopher M. Nguyen, Keith B. Perry, and Natalie L. Denburg. 2010. Pointassist for Older Adults: Analyzing Sub-Movement Characteristics to Aid in Pointing Tasks. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (CHI ’10). ACM, New York, NY, 1115–1124. DOI:
[28]
Kamil K. Imbir, Tomasz Spustek, and Jarosław Żygierewicz. 2016. Effects of Valence and Origin of Emotions in Word Processing Evidenced by Event Related Potential Correlates in a Lexical Decision Task. Frontiers in Psychology 7 (2016), 170040. DOI:
[29]
Giulio Jacucci, Pedram Daee, Tung Vuong, Salvatore Andolina, Khalil Klouche, Mats Sjöberg, Tuukka Ruotsalo, and Samuel Kaski. 2021. Entity Recommendation for Everyday Digital Tasks. ACM Transactions on Computer-Human Interaction 28, 5, Article 29 (2021), 1–41. DOI:
[30]
Jussi P. P. Jokinen, Zhenxin Wang, Sayan Sarcar, Antti Oulasvirta, and Xiangshi Ren. 2020. Adaptive Feature Guidance: Modelling Visual Search with Graphical Layouts. International Journal of Human-Computer Studies 136 (2020), 102376. DOI:
[31]
Antti Kangasrääsiö, Jussi P. P. Jokinen, Antti Oulasvirta, Andrew Howes, and Samuel Kaski. 2019a. Parameter Inference for Computational Cognitive Models with Approximate Bayesian Computation. Cognitive Science 43, 6 (2019), e12738. DOI:. Retrieved from https://onlinelibrary.wiley.com/doi/pdf/10.1111/cogs.12738
[32]
Victor Kaptelinin. 1996. Activity Theory: Implications for Human-Computer Interaction. In Context and Consciousness: Activity Theory and Human-Computer Interaction. B. A. Nardi (Ed.), Vol. 1, The MIT Press, 103–116.
[33]
Saraschandra Karanam and Herre van Oostendorp. 2016. Age-Related Differences in the Content of Search Queries When Reformulating. In Proceedings of the CHI Conference on Human Factors in Computing Systems (CHI ’16). ACM, New York, NY, 5720–5730. DOI:
[34]
Ravi Kumar and Andrew Tomkins. 2010. A Characterization of Online Browsing Behavior. In Proceedings of the 19th International Conference on World Wide Web (WWW ’10). ACM, New York, NY, 561–570. DOI:
[35]
Yuliia Lut, Michael Wang, Elissa M. Redmiles, and Rachel Cummings. 2021. How We Browse: Measurement and Analysis of Digital Behavior. arXiv:2108.06745.
[36]
Shwetambara Malwade, Liezel Cilliers, Xinxin Zhu, Chun-Por Wong, Panagiotis Bamidis, Luis Fernandez-Luque, Yu-Chuan Li, Syed Abdul Shabbir, Mohy Uddin, and Aldilas Achmad Nursetyo. 2018. Mobile and Wearable Technologies in Healthcare for the Ageing Population. Computer Methods and Programs in Biomedicine 161 (Jul. 2018), 233–237. DOI:
[37]
Michael E. Masson. 1982. Cognitive Processes in Skimming Stories. Journal of Experimental Psychology: Learning, Memory, and Cognition 8, 5 (1982), 400.
[38]
Vidhya Navalpakkam and Elizabeth Churchill. 2012. Mouse Tracking: Measuring and Predicting Users’ Experience of Web-Based Content. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (CHI ’12). ACM, New York, NY, 2963–2972. DOI:
[39]
Antti Oulasvirta, Jussi P. P. Jokinen, and Andrew Howes. 2022. Computational Rationality as a Theory of Interaction. In Proceedings of the CHI Conference on Human Factors in Computing Systems (CHI ’22). ACM, New York, NY, Article 359, 14 pages. DOI:
[40]
Chad Peltier and Mark W. Becker. 2017. Individual Differences Predict Low Prevalence Visual Search Performance. Cognitive Research: Principles and Implications 2, 1 (2017), 1–11.
[41]
Matthew Peterson, Melissa Beck, and Jason Wong. 2008. Were You Paying Attention to Where You Looked? The Role of Executive Working Memory in Visual Search. Psychonomic Bulletin & Review 15 (Apr. 2008), 372–377. DOI:
[42]
Benjamin Piwowarski and Hugo Zaragoza. 2007. Predictive User Click Models Based on Click-through History. In Proceedings of the 16th ACM Conference on Conference on Information and Knowledge Management (CIKM ’07). ACM, New York, NY, 175–182. DOI:
[43]
Nilam Ram, Xiao Yang, Mu-Jung Cho, Miriam Brinberg, Fiona Muirhead, Byron Reeves, and Thomas N. Robinson. 2020. Screenomics: A New Approach for Observing and Studying Individuals’ Digital Lives. Journal of Adolescent Research 35, 1 (2020), 16–50. DOI:
[44]
Thomas Redick, Zach Shipstead, Matt Meier, Janelle Montroy, Kenny Hicks, Nash Unsworth, Michael Kane, D. Hambrick, and Randall Engle. 2016. Cognitive Predictors of a Common Multitasking Ability: Contributions from Working Memory, Attention Control, and Fluid Intelligence. Journal of Experimental Psychology General 145, 11 (Jul. 2016), 1473. DOI:
[45]
Yvonne Rogers and Paul Marshall. 2017. Research in the Wild. Synthesis Lectures on Human-Centered Informatics, Vol. 10. Morgan and Claypool Life Sciences, i–97.
[46]
Yvonne Rogers, Nicola Yuill, and Paul Marshall. 2013. Contrasting Lab-Based and In-The-Wild Studies for Evaluating Multi-User Technologies. In: Sara Price, Carey Jewitt and Barry Brown (Eds.), The SAGE Handbook of Digital Technology Research, SAGE, London, 359–373.
[47]
Lena Rosenberg, Anders Kottorp, Bengt Winblad, and Louise Nygård. 2009. Perceived Difficulty in Everyday Technology Use among Older Adults with Or Without Cognitive Deficits. Scandinavian Journal of Occupational Therapy 16, 4 (2009), 216–226. DOI:
[48]
Timothy Salthouse. 2011. What Cognitive Abilities Are Involved in Trail-Making Performance? Intelligence 39 (Jul. 2011), 222–232. DOI:
[49]
Claude Elwood Shannon. 1948. A Mathematical Theory of Communication. The Bell System Technical Journal 27, 3 (1948), 379–423. DOI:
[50]
Joseph Sharit, Mario A. Hernández, Sara J. Czaja, and Peter Pirolli. 2008. Investigating the Roles of Knowledge and Cognitive Abilities in Older Adult Information Seeking on the Web. ACM Transactions on Computer-Human Interaction 15, 1, Article 3 (May 2008), 25 pages. DOI:
[51]
Amanda L. Smith and Barbara S. Chaparro. 2015. Smartphone Text Input Method Performance, Usability, and Preference with Younger and Older Adults. Human Factors 57, 6 (2015), 1015–1028. DOI:
[52]
Ben Steichen, Giuseppe Carenini, and Cristina Conati. 2013. User-Adaptive Information Visualization: Using Eye Gaze Data to Infer Visualization Tasks and User Cognitive Abilities. In Proceedings of the 2013 International Conference on Intelligent User Interfaces (IUI ’13). ACM, New York, NY, 317–328. DOI:
[53]
Dereck Toker, Cristina Conati, Giuseppe Carenini, and Mona Haraty. 2012. Towards Adaptive Information Visualization: On the Influence of User Characteristics. In User Modeling, Adaptation, and Personalization. Judith Masthoff, Bamshad Mobasher, Michel C. Desmarais, and Roger Nkambou (Eds.). Springer, Berlin, 274–285.
[54]
Dereck Toker, Cristina Conati, Ben Steichen, and Giuseppe Carenini. 2013. Individual User Characteristics and Information Visualization: Connecting the Dots through Eye Tracking. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (CHI ’13). ACM, New York, NY, 295–304. DOI:
[55]
Chad Tossell, Philip Kortum, Ahmad Rahmati, Clayton Shepard, and Lin Zhong. 2012. Characterizing Web Use on Smartphones. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (CHI ’12). ACM, New York, NY, 2769–2778. DOI:
[56]
Shari Trewin, John T. Richards, Vicki L. Hanson, David Sloan, Bonnie E. John, Cal Swart, and John C. Thomas. 2012. Understanding the Role of Age and Fluid Intelligence in Information Search. In Proceedings of the 14th International ACM SIGACCESS Conference on Computers and Accessibility (ASSETS ’12). ACM, New York, NY, 119–126. DOI:
[57]
Jodie B. Ullman and Peter M. Bentler. 2012. Structural Equation Modeling. John Wiley & Sons, Ltd. DOI:. Retrieved from https://onlinelibrary.wiley.com/doi/pdf/10.1002/9781118133880.hop202023
[58]
Hiroyuki Umemuro. 2004. Computer Attitudes, Cognitive Abilities, and Technology Usage Among Older Japanese Adults. Gerontechnology 3 (Dec. 2004), 64–76. DOI:
[59]
Nash Unsworth, Richard Heitz, Chad Schrock, and Randall Engle. 2005. An Automated Version of the Operation Span Task. Behavior Research Methods 37 (Sept. 2005), 498–505. DOI:
[60]
Paul van Schaik and Jonathan Ling. 2012. An Experimental Analysis of Experiential and Cognitive Variables in Web Navigation. Human–Computer Interaction 27, 3 (2012), 199–234. DOI:. Retrieved from https://www.tandfonline.com/doi/pdf/10.1080/07370024.2011.646923
[61]
Trish Varao-Sousa, Daniel Smilek, and Alan Kingstone. 2018. In the Lab and in the Wild: How Distraction and Mind Wandering Affect Attention and Memory. Cognitive Research: Principles and Implications 3 (Dec. 2018), 1–9. DOI:
[62]
John Vines, Gary Pritchard, Peter Wright, Patrick Olivier, and Katie Brittain. 2015. An Age-Old Problem: Examining the Discourses of Ageing in HCI and Strategies for Future Research. ACM Transactions on Computer-Human Interaction 22, 1, Article 2 (Feb. 2015), 27 pages. DOI:
[63]
Tung Vuong, Salvatore Andolina, Giulio Jacucci, and Tuukka Ruotsalo. 2021a. Does More Context Help? Effects of Context Window and Application Source on Retrieval Performance. ACM Transactions on Information Systems 40, 2, Article 39 (Sept. 2021), 40 pages. DOI:
[64]
Tung Vuong, Salvatore Andolina, Giulio Jacucci, and Tuukka Ruotsalo. 2021b. Spoken Conversational Context Improves Query Auto-Completion in Web Search. ACM Transactions on Information Systems 39, 3, Article 31 (May 2021), 32 pages. DOI:
[65]
Tung Vuong, Giulio Jacucci, and Tuukka Ruotsalo. 2017a. Proactive Information Retrieval via Screen Surveillance. In Proceedings of the 40th International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR ’17). ACM, New York, NY, 1313–1316. DOI:
[66]
Tung Vuong, Giulio Jacucci, and Tuukka Ruotsalo. 2017b. Watching Inside the Screen: Digital Activity Monitoring for Task Recognition and Proactive Information Retrieval. Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies 1, 3, Article 109 (Sept. 2017), 23 pages. DOI:
[67]
Tung Vuong, Miamaria Saastamoinen, Giulio Jacucci, and Tuukka Ruotsalo. 2019. Understanding User Behavior in Naturalistic Information Search Tasks. Journal of the Association for Information Science and Technology 70, 11 (2019), 1248–1261. DOI:. Retrieved from https://asistdl.onlinelibrary.wiley.com/doi/pdf/10.1002/asi.24201
[68]
Shaun Wallace, Zoya Bylinskii, Jonathan Dobres, Bernard Kerr, Sam Berlow, Rick Treitman, Nirmal Kumawat, Kathleen Arpin, Dave B. Miller, Jeff Huang, and Ben D. Sawyer. 2022. Towards Individuated Reading Experiences: Different Fonts Increase Reading Speed for Different Individuals. ACM Transactions on Computer-Human Interaction 29, 4, Article 38 (Mar. 2022), 56 pages. DOI:
[69]
Jacob O. Wobbrock, Krzysztof Z. Gajos, Shaun K. Kane, and Gregg C. Vanderheiden. 2018. Ability-Based Design. Communications of the ACM 61, 6 (May 2018), 62–71. DOI:
[70]
Ides Y. Wong, Simon S. Smith, and Karen A. Sullivan. 2012. The Relationship between Cognitive Ability, Insight and Self-Regulatory Behaviors: Findings from the Older Driver Population. Accident Analysis & Prevention 49 (2012), 316–321. DOI:
[71]
Yunfeng Zhang and Anthony J. Hornof. 2014. Understanding Multitasking through Parallelized Strategy Exploration and Individualized Cognitive Modeling. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (CHI ’14). ACM, New York, NY, 3885–3894. DOI:
[72]
Xianjun Sam Zheng, Ishani Chakraborty, James Jeng-Weei Lin, and Robert Rauschenberger. 2009. Correlating Low-Level Image Statistics with Users—Rapid Aesthetic and Affective Judgments of Web Pages. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (CHI ’09). ACM, New York, NY, 1–10. DOI:
[73]
Jia Zhou, Pei-Luen Patrick Rau, and Gavriel Salvendy. 2012. Use and Design of Handheld Computers for Older Adults: A Review and Appraisal. International Journal of Human–Computer Interaction 28, 12 (2012), 799–826. DOI:

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Transactions on Computer-Human Interaction
ACM Transactions on Computer-Human Interaction  Volume 31, Issue 3
June 2024
539 pages
EISSN:1557-7325
DOI:10.1145/3613625
Issue’s Table of Contents
This work is licensed under a Creative Commons Attribution International 4.0 License.

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 30 August 2024
Online AM: 07 May 2024
Accepted: 11 March 2024
Revised: 08 March 2024
Received: 17 February 2023
Published in�TOCHI�Volume 31, Issue 3

Check for updates

Author Tags

  1. Cognitive ability
  2. cognitive modeling
  3. digital behavior

Qualifiers

  • Research-article

Funding Sources

  • COADAPT (Human and Work Station Adaptation Supportto aging citizens
  • Academy of Finland (Flagship programme: Finnish Center for Artificial Intelligence FCAI and decision numbers
  • Horizon 2020 FET program of the European Union

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • 0
    Total Citations
  • 297
    Total Downloads
  • Downloads (Last 12 months)297
  • Downloads (Last 6 weeks)147
Reflects downloads up to 21 Oct 2024

Other Metrics

Citations

View Options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Get Access

Login options

Full Access

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media