Abstract
Background: The need to observe brain activity in more natural environments, that is, outside of laboratory settings, is critical for understanding cognition. Wearable low-cost neuroimaging modalities (electroencephalography [EEG] and functional near-infrared spectroscopy [fNIRS]) are portable, noninvasive, and robust to motion artifacts but lack similarly portable tools for use in ecologically valid studies. Smartphones are ubiquitous, programmable, wireless, and thus strong candidates for “pocket laboratories” companion platforms that travel with study subjects. Therefore, we developed the Wearable Cognitive Assessment and Augmentation Toolkit (WearCAAT), a cross-platform neuroimaging task platform that integrates external sensors via the lab streaming layer (LSL) and supports over 100 sensor types. We validated our implementation with healthy human participants under multimodal neuroimaging conditions prior to analysis of data collection in ongoing clinical settings.
Objective: This study aimed to validate WearCAAT, as a platform for functional neuroimaging research, via analysis of human participant data collected during our ongoing National Institutes of Health–funded study.
Methods: We analyzed data from healthy college-aged (ages 18‐30 y) adult participants, who completed a battery of shortened neurocognitive tasks (each lasting 4 min) in WearCAAT, while outfitted with research-grade multimodal EEG and fNIRS sensors. We indicated validity via the presence of task-related behavioral responses and their neuroimaging correlates. As a representative example, we analyzed the visual oddball task due to its well-documented poststimulus features for EEG and fNIRS. We extracted behavioral responses, mean response accuracies, and response times for infrequent (target) and frequent (standard) stimuli classes. We examined, poststimulus, P300, positive amplitude deflection around 300 (ms) in EEG and increased average oxygenated hemoglobin (HbO) levels in fNIRS.
Results: We enrolled a total of 57 (male individuals: n=27, 47%; female individuals: n=30, 53%; mean age 22, SD 3.4 y) participants for data collection. We excluded the first 4 (7%) participants from our analysis due to technical errors. Our analysis revealed increased mean response times for infrequent (target) stimuli (mean 718, SD 148 ms) compared to frequent (standard) stimuli (mean 542, SD 122 ms) with the Wilcoxon test (Z=6.33; P<.001; r=0.87); higher P300 amplitudes over midline regions (parietal and temporal) for EEG; and increased oxygenated hemoglobin over the prefrontal cortex for fNIRS. All participants completed the full battery and reported no usability concerns or app crashes. Similarly, we observed no data loss or corruption that would negatively impact analyses.
Conclusions: WearCAAT-provided outcomes from our study, which analyzed multimodal neuroimaging data collected during a mobile app–based visual oddball task, matched expectations from the literature. While full validation is ongoing for other tasks, we demonstrated initial validity of our app for neurocognitive imaging use. Our app and approach represent the first attempt at dedicated neuroimaging mobile-pocket laboratory and contribute to greater studies in ecological validity.
doi:10.2196/78217
Keywords
Introduction
Functional neuroimaging entails the use of noninvasive brain monitoring sensors during tailored neurocognitive tasks to measure cortical activity in different brain regions corresponding to targeted cognitive domains []. Neuroimaging research offers views into the brain and is critical to the understanding of brain development, injury, and disease or impairment. High-precision techniques such as functional magnetic resonance imaging have limitations that include participant refusal, high cost, and requirements on staying still in the supine position []. This presents an interesting challenge that cognitive neuroscientists have grappled with since the early days of cognitive studies on “ecological validity” [], wherein experimental conditions, particularly those imposed by the nature of laboratory settings and high-precision imaging modalities such as functional magnetic resonance imaging, impact the observed phenomena []. A complementary solution to the challenge of ecological validity is to use wearable, low-power, wireless technologies to combat such challenges facing traditional approaches []. Noninvasive wearable neuroimaging modalities include functional near-infrared spectroscopy (fNIRS) to monitor changes in cerebral hemodynamics related to cognitive activity and electroencephalography [EEG] to monitor neural activations via changes in electrophysiology. Both methods enable observations of human participants in more “natural” environments, and their complementary nature allows overcoming of their individual limitations on temporal and spatial resolution when used together [].
Development and advances in commercially available mobile devices, such as tablets and smartphones, enable portable platforms for conducting human experiments that can further mobilize wearable sensing modalities. The resulting concept “pocket laboratories” describes this paradigm well and poses a unique solution to ecological validity by enabling human participant research in nonstatic environments that could potentially travel with a participant. Researchers in human behavior and medicine have leveraged the widespread adoption of mobile devices as pocket laboratories for more “natural” environments, which are classified under mobile health apps. For example, smartphones and tablets were used in conjunction with neuroimaging devices to measure interactions with websites, apps, and each other, to glean insight into health behaviors such as alcohol consumption [] and Alzheimer disease detection [,]. In addition to the conveniences and benefits that mobile devices offer, they also provide useful hardware for human participant research such as internet connectivity and integrated sensors (ie, gyroscopes, accelerometers, and GPS) [,] and are extensible through Bluetooth connections with external wearable health sensors such as the ones embedded in smartwatches. With an estimated 4.7 billion smartphone users by 2024 [], the potential for participant recruitment is vast, which can further improve not only our understanding of cognition but also allow for the early detection of different conditions, that is, cognitive impairment and monitoring of treatment outcomes through larger studies and populations otherwise unattainable.
Despite the practical uses of mobile devices in clinical research, there’s a notable gap: a platform for conducting general functional neuroimaging research using mobile devices. While accepted tools such as the NIH Toolbox [] developed by the National Institutes of Health provide a platform for gathering psychometrics on human participants and are widely used in clinical settings (225 journals and conferences as of 2022 []), they lack compatibility with functional neuroimaging sensing modalities. Other apps are too limited in scope and lack integration beyond basic proof of concepts and require significant effort to extend to new scenarios. We believe this is because app development is difficult and requires deep technical knowledge and funding that is outside the scope of normal external funding vehicles []. In response to this, we developed a framework for a functional neuroimaging pocket laboratory and provided implementation in the Wearable Cognitive Assessment and Augmentation Toolkit (WearCAAT).
WearCAAT is a cross-platform mobile app, used on both iOS and Android, in conjunction with external single or multimodal sensors, integrated via the lab streaming layer (LSL) []. LSL adds signal synchronization capabilities, equivalent on mobile devices to desktop systems []. However, validation of our framework and implementation is still outstanding. There are numerous challenges in translating a desktop software capability to mobile devices, especially in the domain of functional neuroimaging. Touchscreens are dual-purpose tools that share the responsibility of presenting stimuli and capturing responses via “soft” buttons. Mobile operating systems are sandboxed in nature and typically prevent access to high-precision time-aware clocks, as well as limit multithreading capabilities. A full end-to-end test for our paradigm is necessary to understand the limits and abilities of pocket laboratories in functional neuroimaging.
Methods
Ethical Considerations
Participants signed informed consent before completing cognitive tasks using WearCAAT on an iPad. The consent and collection protocol were reviewed and approved by the Biomedical Research Alliance of New York, LLC (BRANY) [], external institutional review board (1R01AG077018-01), on March 24, 20. Participant privacy was covered under the certificate of confidentiality by the National Institutes of Health that states that researchers will not disclose or use information that may identify participants in any federal, state, or local civil, criminal, administrative, legislative, or other action, suit, or proceeding, even if there is a court subpoena (with exceptions being federal, state, or local law that requires disclosure, or the explicit approval of individual participants to release their name and/or personally identifiable information). Participants were compensated US $20.
Procedure
We examined whether neurocognitive tasks provided by WearCAAT, on commercial mobile devices, reliably elicited cognitive engagement and whether the corresponding biological markers were detectable and identifiable in neuroimaging data. This required that (1) behavioral responses aligned with established task-specific patterns in the literature, and (2) these responses enabled the extraction of physiologically meaningful signals from neuroimaging modalities. Failure to meet both criteria across tasks constituted a negative inconclusive finding, whereas partial success supported the technical validity of our implementation and integration. Consequently, we scoped our analysis on the visual oddball paradigm [], also built in WearCAAT, which is well studied in both EEG and fNIRS modalities for attention monitoring with established expected behavioral, neural, and hemodynamic outcomes [-].
First, we hypothesized that participants’ behavioral data during the visual oddball task, as built in WearCAAT, would exhibit a longer response time (RT) to infrequent (target) stimuli than to frequent (standard) stimuli. Second, we hypothesized that neural correlates for cognition, as measured by EEG and fNIRS during the performance of the visual oddball task, would be detectable in their respective sensing modality when examined using time stamps obtained from the behavioral data. For EEG, the WearCAAT-implemented visual oddball task would evoke a higher P300 subcomponent (positive deflection in amplitude around 300 ms after stimulus) in event-related potentials (ERPs; stimulus-locked activations in EEG) to the infrequent (target) stimuli as compared to the frequent (standard) ones in the midline region. For fNIRS, the average oxygenated hemoglobin (HbO) would positively increase in response to infrequent (target) stimuli as opposed to the frequent (standard) stimuli, in the right prefrontal cortex (PFC).
The remainder of this section describes WearCAAT and the relevant features of this study, followed by the participant information, data collection protocol, the visual oddball task as presented in WearCAAT, and the signal processing pipeline to extract the EEG- and fNIRS-specific components that support or reject our hypotheses.
WearCAAT: An Overview
WearCAAT implements task-based neurocognitive monitoring, wherein participants perform 1 of 11 built-in tasks to elicit known cognitive effects in different domains, including attention, vigilance, working and episodic memory, response inhibition, set shifting, and conflict resolution. Currently implemented tasks in WearCAAT and their cognitive effects are presented in . We built WearCAAT using C# and Extensible Application Markup Language (XAML) from the Multi-App User-Interface (MAUI) framework [] with .NET 8. MAUI provides cross-platform (supporting Android and iOS phones and tablets) app design in a unified project. LSL is integrated using the “slim bindings” approach for high-performance C and C++ libraries to be leveraged through different programming languages, giving direct access to the necessary libraries.
| Task name | Cognitive effect |
| Psychomotor vigilance task [,] | Attention or vigilance |
| Visual oddball paradigm [] | Attention or working memory |
| Go/no-go [] | Response inhibition |
| N-back (n = [0, 1, 2]) [,] | Working memory |
| Stroop [] | Selective attention |
| Flanker [,] | Conflict resolution |
| Wisconsin card sorting [] | Set shifting |
| Verbal memory recognition [] | Episodic memory |
| Bluegrass [] | Episodic memory |
| Resting (eyes = [open, closed]) [] | Default mode net |
Neurocognitive tasks followed the sequence described in . The participant first read the instructions, then started the assessment via tapping the “begin” button, upon which the screen was blank for a configurable resting baseline time window before the task’s logic loop began. The task expired after the set amount of time and was followed by a second baseline period. Timing information was determined using a stopwatch object, which counted monotonically from the start, as is common in psychometrics platforms []. Tasks were configurable in the app to allow further flexibility and experimentation, specifically regarding stimulus type, timing, interstimulus interval information, stimulus presentation ratios, and more, depending on the task.

Task events were broken down into three categories as follows: (1) metadata pertaining to task information and configuration for auditing purposes; (2) stimulus event appearances, types, etc; and (3) user markers, button presses, or other responses. Each category streamed data to a corresponding LSL stream outlet, which sent the data wirelessly to the recording platform. depicts the sequence of events and the respective streams for each task.
Participants
We recruited 57 (male individuals: n=27, 47%; female individuals: n=30, 53%) participants, aged 18 to 30 (mean age 22, SD 3.4 y) years, from the undergraduate and graduate student bodies at Villanova University via flyers posted in common university spaces. We detail participant demographic data in . Exclusion criteria included current or past severe neurological or psychiatric disorders and significant vision or hearing impairments. Participants first attended an initial screening, where we collected demographic data and relevant medical histories via a written survey, measured the participant’s head circumference to determine neuroimaging device cap size, and scheduled a follow-up data collection session. After participants signed the informed consent form, we collected behavioral and neuroimaging data from them while they used WearCAAT in a session that took approximately 1.5 hours.
| Demographics | Male (n=27), n (%) | Female (n=30), n (%) | Total (N=57), n (%) |
| Asian | 8 (30) | 7 (23) | 15 (26) |
| Black or African American | 1 (4) | 0 (0) | 1 (2) |
| Hispanic | 2 (7) | 5 (17) | 7 (12) |
| White | 13 (48) | 13 (43) | 26 (46) |
| White and Black or African American | 2 (7) | 0 (0) | 2 (4) |
| White and Hispanic | 1 (4) | 5 (17) | 6 (11) |
Data Collection
We administered all 11 abbreviated tasks to participants via WearCAAT in sequence, where the first 2 were the resting tasks (eyes opened and eyes closed) and the remaining 9 tasks were presented in randomized order (). All participants completed the entire task battery built in WearCAAT in 1 sitting while wearing a full head cap housing the multiple optodes. The electrodes formed a combined wireless fNIRS-EEG system (NIRSport2, NIRx Medizintechnik GmbH and Smarting-mbt wireless EEG, mBrainTrain, respectively) [,]. Using our hybrid neuroimaging system, we collected 51-channel fNIRS (43 long and 8 short distances) and 32-channel EEG data from the frontal, temporal, and parietal regions of the brain simultaneously. We used 3 distinct flexible cap sizes as provided by NIRx company []—small (56 cm), medium (58 cm), and large (60 cm)—all with 128 slits for probe locations identified according to the 10-20 international system to accommodate for different head sizes, improve comfort, and ensure fNIRS and EEG measurements with good coupling from similar head locations. The complete layout of our protocol is detailed in ; the LSL data streams and their respective interfaces are all coordinated through a Wireless Area Network (WAN) hosted on a private router with no external internet connection nor devices.
| ROI | EEG channels | fNIRS channels |
| Frontal left |
|
|
| Frontal right |
|
|
| Temporal left |
|
|
| Temporal right |
|
|
| Parietal left |
|
|
| Parietal right |
|
|
| Frontal midline |
| — |
| Parietal midline |
| — |
| Temporal midline |
| — |
aNot available.

Behavioral responses were recorded in WearCAAT where task events (eg, stimulus presentation times, user responses via button presses, etc) were time stamped in WearCAAT and wirelessly streamed through LSL to a laptop (Windows 10) running App-LabRecorder []. Concurrent EEG and fNIRS data were also streamed wirelessly to the same laptop, which synchronized all clock times and removed jitter automatically. While WearCAAT supports Android, we only used an Apple iPad Pro (sixth generation) as our mobile platform due to the consistency of iOS devices. Using different operating systems and devices from different manufacturers might introduce errors that are harder to quantify [], and that is out of scope for this body of work.
Evaluation Protocol: Visual Oddball Paradigm
To provide empirical support and initial validation, we reported our neuroimaging and behavioral outcomes obtained from the abbreviated visual oddball task implemented via WearCAAT. In this task, participants were presented with either 1 of 2 types of visual stimuli where each consisted of 5 repeated letters and asked to respond by tapping 1 of the 2, parallel and equally sized, buttons at the bottom of the screen with the index finger on their dominant hand. The target stimulus (“XXXXX”) matched with the left-most button labeled “TARGET,” and the standard stimulus (“OOOOO”) matched to the right-most button. The interstimulus interval was 2 seconds, with stimuli presentations lasting 0.5 seconds and the screen remaining blank for 1.5 seconds. Target stimuli appeared infrequently relative to the standard, with at least 7 to 21 standard stimuli appearing between each target presentation to reduce participant expectation. On average, each participant witnessed 11.87 (SD 1.15) standard stimuli between each successive target stimulus. The total duration of the task was 4 minutes and occurred between two 10-second baseline periods. Participants received instructions verbally from the experimenters and in the app before beginning each task, as depicted in . Participants began the task by tapping the “begin” button, which started the baseline period followed by the oddball sequence.

Signal Processing
Overview
The simultaneous 51-channel fNIRS and 32-channel EEG data were collected from the full head in frontal, temporal, and parietal locations as shown in , with regions of interest (ROI) detailed in . Our fNIRS and EEG data processing pipeline for artifact removal (motion, physiological, environmental, or equipment noise) and data conversion (hemodynamic response extraction) was performed offline using custom-built MATLAB (version R2024; MathWorks, Inc) codes [] and in accordance with published best practices [-]. We used the 8 fNIRS short channels to remove skin artifacts. EEG data were additionally processed for the removal of eye blinks, eye movement, muscle artifacts, power line noise, and limiting the data within the range of 0 Hz to 45 Hz.

Given the large number of channels (32 for EEG and 43 long separations for fNIRS), we organized the data into ROIs based on both anatomical and functional considerations (detailed in ). To manage the data efficiently, we averaged the responses for each stimulus (target and standard) separately across all 57 participants, focusing on “valid” data from each participant, channel, and stimulus. We determined a channel to be invalid based on impedance for EEG (>30 kΩ) [], and Scalp Coupling Index (<0.4) for fNIRS []. Finally, we averaged the score per channel across all participants and grouped the channels based on their respective ROIs, reducing the dimensionality of the dataset.
WearCAAT Markers: User Behavioral Responses
We gathered the response accuracy by evaluating the correct and incorrect user responses and extracting the reaction times to target and standard stimuli. We determined significance between target and standard responses using the Wilcoxon test, a nonparametric alternative to the paired 2-tailed t test when the data do not follow a normal distribution.
EEG: Event-Related Potential
We extracted stimulus-locked data epochs (ERPs) from EEG recordings using 0.2-second prestimulus and 1-second poststimulus interval and performed a baseline correction by subtracting the mean of the prestimulus data from the whole data epoch. We then removed intrinsic response variability by averaging multiple trials of epochs within the task to obtain an averaged ERP for each stimulus (target and standard), separately. Finally, we gathered the ERP features by extracting the positive and negative peak amplitudes with their respective timings, specifically focusing on the P300 component (positive deflection around 300 ms after the stimulus), which was shown to be higher in target response as compared to the standard one in the visual oddball task in healthy young adults [,].
fNIRS: Oxygenated Hemoglobin
We processed our fNIRS data by removing motion artifacts and baseline shifts through wavelet and spline filters []. We further removed physiological signals, cardiac, respiratory, and Mayer waves, with a finite impulse response low-pass filter using the cutoff frequency 0.08 Hz []. Finally, we converted the light intensity measurements into changes in HbO and deoxygenated hemoglobin using the modified Beer-Lambert Law with published coefficients from the literature []. After using short channel recordings to remove potential skin blood flow artifacts from long channel recordings using a general linear model [], we extracted 20-second poststimulus data epochs and applied baseline correction using the 1-second prestimuli onset. Notably, since HbO was the most used fNIRS measure in studies implementing the visual oddball task that was indicative of cognitive activity–related changes in attention domain–specific apps [], we focused our results and comparisons to only HbO outcomes in this study.
Results
Overview
We enrolled 57 participants in data collection, and all of them performed all 11 tasks in 1 sitting. We observed zero participant dropout with no app crashes or corrupted data. We excluded data from our first 4 (7%) participants; 2 due to poor impedance from improper cap setup, and 2 after a patch to WearCAAT that altered the timing logic to improve the responsiveness of the touch screen during timed loops. WearCAAT collected and synchronized multiple concurrent streams of task-related data (1 stream for stimulus; 1 for each button press; and 1 for metadata and task events, such as task start and stop and baseline start and stop) with no data loss in participant responses or disconnects from the recording server. We also observed no additional loss of information or signal content from the combined fNIRS-EEG sensors as well. Participants reported no complaints or concerns with the WearCAAT app, the instructions provided either in app or verbally, or the overall data collection procedure, indicating a low participant burden. An example participant can be seen sitting comfortably during collection in .

User Response Times
We found that the mean RT for target stimuli was 718 (SD 148) milliseconds and, for standard stimuli, it was 542 (SD 122) milliseconds. The Wilcoxon test revealed significant differences in RT to target (infrequent) stimuli as compared to the standard (frequent) ones (Z=6.33; P<.001; r=0.87). These outcomes indicated that the participants took longer to identify the target stimulus than the standard stimulus. Participants’ mean percent accuracy for identifying stimuli was 94.58 (SD 7.412) for the target stimulus and 99.12 (SD 1.447) for the standard stimulus, suggesting that participants identified and responded to the frequent stimulus more correctly as compared to the infrequent ones overall.
EEG: Extracted P300
All participant-averaged (SEM) ERP waveforms obtained using WearCAAT in an iPad for the abbreviated visual oddball task for target (red) and standard (blue) stimulus in the parietal midline (Pz) and central midline (Cz) regions are presented in , separately. Our results showed higher P300 amplitude in response to target stimuli as compared to the standard one, especially in the midline regions on the Cz and Pz locations, in line with the published literature on computerized presentation of a regular length (approximately 20 min) visual oddball task. The Cz, presented in , recorded a positive peak within the 250-millisecond to 400-millisecond intervals at 356 milliseconds, having an amplitude of 3.5397 (SD 0.6258) μV for the target stimulus and at 372 milliseconds with an amplitude of 0.74667 (SD 0.4929) μV for the standard stimulus.

Similarly, the Pz, presented in , recorded a peak at 296 ms with an amplitude of 2.7219 (SD 0.4756) μV in the target stimulus and a peak at 268 ms with an amplitude of 0.949 (SD 0.3637) μV in the standard stimulus.

fNIRS: Trends in Oxygenated Hemoglobin
As presented in , the overall participant-averaged (SEM) HbO activations in the right PFC demonstrated a clear, positive increase in response to the target (red) stimulus peaking at approximately 9 seconds and then returning to lower values following the shape of a typical hemodynamic response function. The standard stimulus (blue) did not generate an increased HbO pattern.

Same trends on HbO in response to target and standard stimuli were observed in all right frontal fNIRS channels constituting the right frontal ROI, as also presented separately in .
Discussion
Principal Findings
We presented our analysis for the abbreviated visual oddball task as presented and collected using WearCAAT and our procedures. We found indications that electrophysiological and hemodynamic activation patterns for the brain observed with simultaneously collected fNIRS and EEG data follow expected trends, despite the shorter runtime (4 min as opposed to 20 min in commonly implemented versions of the task) and the mobile app platform (as opposed to a computer screen).
Our initial hypothesis for the behavioral responses was supported. We observed greater RTs to target (infrequent) versus standard (frequent) stimuli. The Wilcoxon signed-rank test demonstrated the significance for RT between infrequent and frequent stimuli. These findings are consistent with previously reported values in the literature for the visual oddball task [,].
Interestingly, we also observed that periods of responses for target stimuli were greater than some reported values, whereas the standard stimuli RTs were much closer. This may be attributed, in part, to the physical differences in iPad “soft” buttons and the typical hardware switches commonly used with the visual oddball task []. Traditional desktop setups report participants using 2 distinct controllers (1 per hand), which dedicate a controller response per stimulus type. In contrast with our study, participants used a singular index finger to switch between button presses. The typical delay reported between stimulus types could be exacerbated by physical delays introduced by a participant needing to move his finger from hovering over 1 button to another one on the opposite side of the iPad screen. As future work, we will ensure that participants are instructed on how to hold the iPad, with relevant findings from the literature.
Our second hypothesis regarding EEG was supported by the successful extraction of P300 subcomponents from the ERP waveform, described as a positive deflection in amplitude in response to the target stimulus, appearing around 300 milliseconds after the stimulus from our collected data. We obtained higher P300 amplitude in the midline ROIs (Pz and Cz) for the infrequent (target) stimuli as compared to the frequent (standard) stimuli, following the expected outcome as reported in the literature [,,,]. Our average latencies for P300 peaks in Pz and Cz were also within previously reported bounds [].
We noted the visual jaggedness of the P300 signals, seen in , which we expected to be smoother, as reported in the literature. This could be caused by a combination of the shortened task times and the grand averaging technique used for analysis. Typical studies report task lengths of 20 minutes or greater for the visual oddball task, whereas this study’s task length was 4 minutes. Longer task times would produce 5 times more trials for both target and standard stimuli per participant, the averages of which would smooth out irregularities and potential physiological artifacts in the time series. Further study using WearCAAT, EEG, and the visual oddball task with longer task times would provide more clarity on the matter.
Our second hypothesis regarding fNIRS was also supported by the positive average HbO increases measured in the right PFC in response to infrequent (target) stimuli as compared to the frequent (standard) stimuli. Specifically, we observed the increases in HbO in the frontal right ROI, which was widely reported in the fNIRS literature where computerized and traditional length visual oddball task was used [,]. In fact, such findings were prominent in all right frontal channels when considered separately as well as demonstrating an attention domain–specific global activation in right PFC as measured by fNIRS.
On usability, we point to the smoothness of data collection. Specifically, the lack of complaints from the participants and experimenters, combined with the zero-dropout rate and app crashes, is to be noted. Given that users’ major concern with mobile health apps is the perceived bugginess and clunkiness of apps [], we incorporated haptics and button color changes as feedback to users’ actions. We assume the responsiveness and perceived functionality of our WearCAAT implementation is tolerable to young adults who are most fluent and comfortable in app use. However, because we did not perform a formal qualitative post–data collection survey, our interpretation is limited to “no complaints were reported.” This limitation ought to be accounted for in future studies with formal participant surveys after participation to add qualitative metrics for perceived clunkiness and usability.
Conclusions
In this study, we provided evidence for the technical validation of mobile devices in task-based functional neuroimaging research via the analysis of multimodal EEG-fNIRS and behavioral data collected during an abbreviated mobile visual oddball task from 57 healthy young adults. Specifically, our goal was to evaluate whether behavioral effects, higher mean responses to infrequent (target) versus frequent (standard) stimuli, were present across participants. We also determined if the P300 component obtained from the ERP waveform on the midline and increases in measured HbO over the right PFC, as measured by fNIRS for target stimuli as compared to the standard ones, can be simultaneously captured using the visual oddball task as implemented in our mobile app WearCAAT. All desired features were elicited using an abbreviated visual oddball task on a mobile platform, which demonstrated the validity of WearCAAT functionality and synchrony for functional neuroimaging studies.
While future work entails the validation of more tasks implemented in the current iteration of WearCAAT, and comparisons of fNIRS and EEG features for young versus older adults, this work supports the use of mobile platforms for cognitive neuroimaging.
WearCAAT will soon be easily accessible through both Google Play and Apple App Stores. It is our hope that the wide range of reconfigurable neurocognitive tasks, usability, and ease of use with extant neuroimaging setups will enable nontechnical users to leverage mobile pocket laboratories in future studies and begin to answer outstanding questions in ecological validity. We believe the validation of technical ability as reported in this experiment lends confidence to the pocket lab paradigm and informs future studies into human behavior, in and out in the wild.
Funding
This research was supported by a grant from the National Institutes of Health (grant R01AG077018) awarded to MI. The funding source has no role in the design, data collection and analysis, interpretation of the study, writing, or decision to submit for publication.
Authors' Contributions
PR contributed to app and task designs, app development, writing the original draft, writing, reviewing, and editing. MI contributed to the study, task and app designs, data analysis, interpretation, writing, reviewing, and editing. LG contributed to data collection, data analysis and interpretation, writing, reviewing, and editing. RH contributed to study and task designs, writing, reviewing, and editing.
Conflicts of Interest
None declared.
References
- Khanna N, Altmeyer W, Zhuo J, Steven A. Functional neuroimaging: fundamental principles and clinical applications. Neuroradiol J. Apr 2015;28(2):87-96. [CrossRef] [Medline]
- Holtzer R, Epstein N, Mahoney JR, Izzetoglu M, Blumen HM. Neuroimaging of mobility in aging: a targeted review. J Gerontol A Biol Sci Med Sci. Nov 2014;69(11):1375-1388. [CrossRef] [Medline]
- Neisser U. Cognition and Reality: Principles and Implications of Cognitive Psychology. W.H. Freeman; 1976. URL: http://archive.org/details/cognitionreality00neis [Accessed 2024-05-01]
- Bronfenbrenner U. Toward an experimental ecology of human development. American Psychologist. 1977;32(7):513-531. [CrossRef]
- von Lühmann A, Zheng Y, Ortega-Martinez A, et al. Towards Neuroscience of the Everyday World (NEW) using functional near-infrared spectroscopy. Curr Opin Biomed Eng. Jun 2021;18:100272. [CrossRef] [Medline]
- Jungnickel E, Gramann K. Mobile Brain/Body Imaging (MoBI) of physical interaction with dynamically moving objects. Front Hum Neurosci. 2016;10:306. [CrossRef] [Medline]
- Wu T, Sherman G, Giorgi S, et al. Smartphone sensor data estimate alcohol craving in a cohort of patients with alcohol-associated liver disease and alcohol use disorder. Hepatol Commun. Dec 1, 2023;7(12):12. [CrossRef] [Medline]
- Coughlan G, Coutrot A, Khondoker M, Minihane AM, Spiers H, Hornberger M. Toward personalized cognitive diagnostics of at-genetic-risk Alzheimer’s disease. Proc Natl Acad Sci U S A. May 7, 2019;116(19):9285-9292. [CrossRef] [Medline]
- Taylor JC, Heuer HW, Clark AL, et al. Feasibility and acceptability of remote smartphone cognitive testing in frontotemporal dementia research. Alzheimers Dement (Amst). 2023;15(2):e12423. [CrossRef] [Medline]
- Hackett K, Xu S, McKniff M, Paglia L, Barnett I, Giovannetti T. Mobility-based smartphone digital phenotypes for unobtrusively capturing everyday cognition, mood, and community life-space in older adults: feasibility, acceptability, and preliminary validity study. JMIR Hum Factors. Nov 22, 2024;11:e59974. [CrossRef] [Medline]
- Zawada S, Acosta J, Collins C, et al. Real-world smartphone data predicts mood after ischemic stroke and transient ischemic attack symptoms and may constitute digital endpoints: a proof-of-concept study. Mayo Clin Proc Digit Health. Sep 2025;3(3):100240. [CrossRef] [Medline]
- Degenhard J. Number of smartphone users worldwide from 2014 to 2029. Statista. URL: https://www.statista.com/forecasts/1143723/smartphone-users-in-the-world [Accessed 2024-05-16]
- Hodes RJ, Insel TR, Landis SC, NIH Blueprint for Neuroscience Research. The NIH toolbox: setting a standard for biomedical research. Neurology (ECronicon). Mar 12, 2013;80(11 Suppl 3):S1. [CrossRef] [Medline]
- Fox RS, Zhang M, Amagai S, et al. Uses of the NIH Toolbox® in clinical samples. Neur Clin Pract. Aug 2022;12(4):307-319. [CrossRef]
- Siegler AJ, Knox J, Bauermeister JA, Golinkoff J, Hightow-Weidman L, Scott H. Mobile app development in health research: pitfalls and solutions. Mhealth. 2021;7:32. [CrossRef] [Medline]
- Kothe C, Shirazi SY, Stenner T, et al. The lab streaming layer for synchronized multimodal recording. Neuroscience. Preprint posted online on Jul 14, 2024. [CrossRef]
- Blum S, Hölle D, Bleichner MG, Debener S. Pocketable labs for everyone: synchronized multi-sensor data streaming and recording on smartphones with the lab streaming layer. Sensors (Basel). Dec 5, 2021;21(23):8135. [CrossRef] [Medline]
- BRANY. URL: https://www.brany.com/ [Accessed 2024-11-27]
- McCarthy G, Luby M, Gore J, Goldman-Rakic P. Infrequent events transiently activate human prefrontal and parietal cortex as measured by functional MRI. J Neurophysiol. Mar 1997;77(3):1630-1634. [CrossRef] [Medline]
- Izzetoglu M, Bunce SC, Izzetoglu K, Onaral B, Pourrezaei AK. Functional brain imaging using near-infrared technology. IEEE Eng Med Biol Mag. 2007;26(4):38-46. [CrossRef] [Medline]
- Patel SH, Azzam PN. Characterization of N200 and P300: selected studies of the event-related potential. Int J Med Sci. 2005;2(4):147-154. [CrossRef] [Medline]
- Ardekani BA, Choi SJ, Hossein-Zadeh GA, et al. Functional magnetic resonance imaging of brain activity in the visual oddball task. Cognitive Brain Research. Nov 2002;14(3):347-356. [CrossRef]
- Rizer W, Aday JS, Carlson JM. Changes in prefrontal cortex near infrared spectroscopy activity as a function of difficulty in a visual P300 paradigm. J Near Infrared Spectrosc. Aug 2018;26(4):222-228. [CrossRef]
- Classon I. Migrating From Xamarin.Forms to .NET MAUI: A Comprehensive Guide. Apress; 2025. [CrossRef] ISBN: 9798868812156
- Kay M, Rector K, Consolvo S, et al. PVT-touch: adapting a reaction time test for touchscreen devices. IEEE Presented at: ICTs for improving Patients Rehabilitation Research Techniques; May 5-8, 2013; Venice, Italy. [CrossRef]
- Evans MS, Harborne D, Smith AP. Developing an objective indicator of fatigue: an alternative mobile version of the psychomotor vigilance task (m-PVT). In: Longo L, Leva MC, editors. Human Mental Workload: Models and Applications. Springer; 2019:49-71. [CrossRef] ISBN: 9783030142735
- Verbruggen F, Logan GD. Automatic and controlled response inhibition: associative learning in the go/no-go and stop-signal paradigms. J Exp Psychol Gen. Nov 2008;137(4):649-672. [CrossRef] [Medline]
- Kane MJ, Conway ARA, Miura TK, Colflesh GJH. Working memory, attention control, and the N-back task: a question of construct validity. J Exp Psychol Learn Mem Cogn. May 2007;33(3):615-622. [CrossRef] [Medline]
- Owen AM, McMillan KM, Laird AR, Bullmore E. N-back working memory paradigm: a meta-analysis of normative functional neuroimaging studies. Hum Brain Mapp. May 2005;25(1):46-59. [CrossRef] [Medline]
- Spanakis P, Jones A, Field M, Christiansen P. A Stroop in the hand is worth two on the laptop: superior reliability of a smartphone based alcohol Stroop in the real world. Subst Use Misuse. 2019;54(4):692-698. [CrossRef] [Medline]
- Yeung MK, Lee TL, Chan AS. Neurocognitive development of Flanker and Stroop interference control: a near-infrared spectroscopy study. Brain Cogn. Aug 2020;143:105585. [CrossRef] [Medline]
- Pronk T, Hirst RJ, Wiers RW, Murre JMJ. Can we measure individual differences in cognitive measures reliably via smartphones? A comparison of the flanker effect across device types and samples. Behav Res Methods. Jun 2023;55(4):1641-1652. [CrossRef] [Medline]
- Nyhus E, Barceló F. The Wisconsin Card Sorting Test and the cognitive assessment of prefrontal executive functions: a critical update. Brain Cogn. Dec 2009;71(3):437-451. [CrossRef] [Medline]
- Wang LM, Huang YH, Chou PH, Wang YM, Chen CM, Sun CW. Characteristics of brain connectivity during verbal fluency test: convolutional neural network for functional near-infrared spectroscopy analysis. J Biophotonics. Jan 2022;15(1):1. [CrossRef] [Medline]
- Broster LS, Li J, Wagner B, et al. Spared behavioral repetition effects in Alzheimer’s disease linked to an altered neural mechanism at posterior cortex. J Clin Exp Neuropsychol. Sep 14, 2018;40(8):761-776. [CrossRef]
- Nguyen T, Babawale O, Kim T, Jo HJ, Liu H, Kim JG. Exploring brain functional connectivity in rest and sleep states: a fNIRS study. Sci Rep. Nov 2018;8(1):16144. [CrossRef]
- BioRender. URL: https://BioRender.com [Accessed 2026-01-26]
- Mobile EEG-fNIRS – LiveAmp with actiCAP electrodes and NIRx NIRSport2. Brain Products. 2020. URL: https://pressrelease.brainproducts.com/eeg-fnirs-setup/ [Accessed 2024-01-19]
- Concurrent fNIRS and EEG. NIRx Medical Technologies. URL: https://nirx.net/fnirs-eeg [Accessed 2024-01-18]
- Customized NIRScaps and probes. NIRx Medical Technologies. URL: https://nirx.net/nirscap [Accessed 2025-05-05]
- Kothe C. Labstreaminglayer/app-labrecorder. GitHub. May 9, 2024. URL: https://github.com/labstreaminglayer/App-LabRecorder [Accessed 2024-05-15]
- Woods D, Yund W, Pebler P, Grivich MI. The timing precision of iOS and Android apps. Neurobehavioral Systems, Inc. URL: https://cdn1.neurobs.com/misc/mobile_timing_poster.pdf [Accessed 2026-01-19]
- MATLAB version: 9.13.0 (r2022b). The MathWorks Inc. URL: https://www.mathworks.com [Accessed 2025-05-15]
- Huang R, Hong KS, Yang D, Huang G. Motion artifacts removal and evaluation techniques for functional near-infrared spectroscopy signals: a review. Front Neurosci. 2022;16:878750. [CrossRef] [Medline]
- Abdelnour AF, Huppert T. Real-time imaging of human brain function by near-infrared spectroscopy using an adaptive general linear model. Neuroimage. May 15, 2009;46(1):133-143. [CrossRef] [Medline]
- Li R, Yang D, Fang F, Hong KS, Reiss AL, Zhang Y. Concurrent fNIRS and EEG for brain function investigation: a systematic, methodology-focused review. Sensors (Basel). Aug 2022;22(15):5865. [CrossRef]
- Habibzadeh Tonekabony Shad E, Molinas M, Ytterdal T. Impedance and noise of passive and active dry EEG electrodes: a review. IEEE Sensors J. Dec 2020;20(24):14565-14577. [CrossRef]
- Pollonini L, Bortfeld H, Oghalai JS. PHOEBE: a method for real time mapping of optodes-scalp coupling in functional near-infrared spectroscopy. Biomed Opt Express. Dec 1, 2016;7(12):5104-5119. [CrossRef] [Medline]
- Luck SJ. An Introduction to the Event-Related Potential Technique. 2nd ed. MIT Press; 2014. ISBN: 9780262621960
- Delorne A, Makeig S. Extract data epochs. EEGLAB Wiki. URL: https://eeglab.org/tutorials/07_Extract_epochs/Extracting_Data_Epochs.html [Accessed 2024-06-27]
- Yücel MA, Selb J, Aasted CM, et al. Mayer waves reduce the accuracy of estimated hemodynamic response functions in functional near-infrared spectroscopy. Biomed Opt Express. 2016;7(8):3078. [CrossRef]
- Bunce SC, Izzetoglu M, Izzetoglu K, Onaral B, Pourrezaei K. Functional near-infrared spectroscopy. IEEE Eng Med Biol Mag. 2006;25(4):54-62. [CrossRef] [Medline]
- Uga M, Dan I, Sano T, Dan H, Watanabe E. Optimizing the general linear model for functional near-infrared spectroscopy: an adaptive hemodynamic response function approach. Neurophotonics. Jul 2014;1(1):015004. [CrossRef] [Medline]
- Polich J, Bondurant T. P300 sequence effects, probability, and interstimulus interval. Physiol Behav. Jun 1997;61(6):843-849. [CrossRef] [Medline]
- Polich J. Updating P300: an integrative theory of P3a and P3b. Clin Neurophysiol. Oct 2007;118(10):2128-2148. [CrossRef] [Medline]
- Muñoz V, Muñoz-Caracuel M, Angulo-Ruiz BY, Gómez CM. Neurovascular coupling during auditory stimulation: event-related potentials and fNIRS hemodynamic. Brain Struct Funct. Nov 2023;228(8):1943-1961. [CrossRef] [Medline]
- Poskotinova L, Khasanova N, Kharak A, Krivonogova O, Krivonogova E. Parameters of auditory evoked related potentials p300 in disorders of different cognitive function domains (visuospatial/executive and memory) in elderly hypertensive persons. Diagnostics (Basel). Apr 30, 2023;13(9):1598. [CrossRef] [Medline]
Abbreviations
| BRANY: Biomedical Research Alliance of New York |
| Cz: central midline |
| EEG: electroencephalogram |
| ERP: event-related potential |
| fNIRS: functional near-infrared spectroscopy |
| HbO: oxygenated hemoglobin |
| LSL: lab streaming layer |
| PFC: prefrontal cortex |
| Pz: parietal midline |
| ROI: region of interest |
| RT: response time |
| WearCAAT: Wearable Cognitive Assessment and Augmentation Toolkit |
Edited by Alessandro Zampogna; submitted 28.May.2025; peer-reviewed by Martina Patera, Stephanie Zawada; final revised version received 31.Oct.2025; accepted 11.Nov.2025; published 06.Feb.2026.
Copyright© Peter Rokowski, Meltem Izzetoglu, Luis Gomero, Roee Holtzer. Originally published in JMIR Neurotechnology (https://neuro.jmir.org), 6.Feb.2026.
This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Neurotechnology, is properly cited. The complete bibliographic information, a link to the original publication on https://neuro.jmir.org, as well as this copyright and license information must be included.

