<?xml version="1.0" encoding="UTF-8"?><!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.0 20040830//EN" "journalpublishing.dtd"><article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" dtd-version="2.0" xml:lang="en" article-type="research-article"><front><journal-meta><journal-id journal-id-type="nlm-ta">JMIR Serious Games</journal-id><journal-id journal-id-type="publisher-id">games</journal-id><journal-id journal-id-type="index">15</journal-id><journal-title>JMIR Serious Games</journal-title><abbrev-journal-title>JMIR Serious Games</abbrev-journal-title><issn pub-type="epub">2291-9279</issn><publisher><publisher-name>JMIR Publications</publisher-name><publisher-loc>Toronto, Canada</publisher-loc></publisher></journal-meta><article-meta><article-id pub-id-type="publisher-id">v14i1e83128</article-id><article-id pub-id-type="doi">10.2196/83128</article-id><article-categories><subj-group subj-group-type="heading"><subject>Original Paper</subject></subj-group></article-categories><title-group><article-title>Analysis of Spatiotemporal Features in a Virtual Navigation Game Across Different Age Groups: Quantitative Research</article-title></title-group><contrib-group><contrib contrib-type="author"><name name-style="western"><surname>Qiao</surname><given-names>Xiaofeng</given-names></name><degrees>BEng</degrees><xref ref-type="aff" rid="aff1">1</xref><xref ref-type="aff" rid="aff2">2</xref><xref ref-type="aff" rid="aff3">3</xref><xref ref-type="aff" rid="aff4">4</xref></contrib><contrib contrib-type="author" equal-contrib="yes"><name name-style="western"><surname>Tian</surname><given-names>Shan</given-names></name><degrees>PhD</degrees><xref ref-type="aff" rid="aff1">1</xref><xref ref-type="aff" rid="aff2">2</xref><xref ref-type="aff" rid="aff3">3</xref><xref ref-type="aff" rid="aff4">4</xref><xref ref-type="fn" rid="equal-contrib1">*</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Tang</surname><given-names>Min</given-names></name><degrees>MEng</degrees><xref ref-type="aff" rid="aff1">1</xref><xref ref-type="aff" rid="aff2">2</xref><xref ref-type="aff" rid="aff3">3</xref><xref ref-type="aff" rid="aff4">4</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>He</surname><given-names>Shipei</given-names></name><degrees>MSc</degrees><xref ref-type="aff" rid="aff1">1</xref><xref ref-type="aff" rid="aff2">2</xref><xref ref-type="aff" rid="aff3">3</xref><xref ref-type="aff" rid="aff4">4</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Wang</surname><given-names>Jinghui</given-names></name><degrees>MEng</degrees><xref ref-type="aff" rid="aff1">1</xref><xref ref-type="aff" rid="aff2">2</xref><xref ref-type="aff" rid="aff3">3</xref><xref ref-type="aff" rid="aff4">4</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Fan</surname><given-names>Linyuan</given-names></name><degrees>BEng</degrees><xref ref-type="aff" rid="aff1">1</xref><xref ref-type="aff" rid="aff2">2</xref><xref ref-type="aff" rid="aff3">3</xref><xref ref-type="aff" rid="aff4">4</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Zhu</surname><given-names>Yuanjie</given-names></name><degrees>MEng</degrees><xref ref-type="aff" rid="aff1">1</xref><xref ref-type="aff" rid="aff2">2</xref><xref ref-type="aff" rid="aff3">3</xref><xref ref-type="aff" rid="aff4">4</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Zhang</surname><given-names>Zhiyang</given-names></name><degrees>MD</degrees><xref ref-type="aff" rid="aff1">1</xref><xref ref-type="aff" rid="aff2">2</xref><xref ref-type="aff" rid="aff3">3</xref><xref ref-type="aff" rid="aff4">4</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Du</surname><given-names>Songjun</given-names></name><degrees>BEng</degrees><xref ref-type="aff" rid="aff1">1</xref><xref ref-type="aff" rid="aff2">2</xref><xref ref-type="aff" rid="aff3">3</xref><xref ref-type="aff" rid="aff4">4</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Dong</surname><given-names>Chaojie</given-names></name><degrees>BEng</degrees><xref ref-type="aff" rid="aff1">1</xref><xref ref-type="aff" rid="aff2">2</xref><xref ref-type="aff" rid="aff3">3</xref><xref ref-type="aff" rid="aff4">4</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Chen</surname><given-names>Yepu</given-names></name><degrees>BD</degrees><xref ref-type="aff" rid="aff1">1</xref><xref ref-type="aff" rid="aff2">2</xref><xref ref-type="aff" rid="aff3">3</xref><xref ref-type="aff" rid="aff4">4</xref></contrib><contrib contrib-type="author" corresp="yes" equal-contrib="yes"><name name-style="western"><surname>Liu</surname><given-names>Xiaoyu</given-names></name><degrees>PhD</degrees><xref ref-type="aff" rid="aff1">1</xref><xref ref-type="aff" rid="aff2">2</xref><xref ref-type="aff" rid="aff3">3</xref><xref ref-type="aff" rid="aff4">4</xref><xref ref-type="fn" rid="equal-contrib1">*</xref></contrib></contrib-group><aff id="aff1"><institution>School of Biological Science and Medical Engineering, Beihang University</institution><addr-line>5 Building in Beihang University, 37 Xueyuan Road</addr-line><addr-line>Beijing</addr-line><country>China</country></aff><aff id="aff2"><institution>Key Laboratory of Biomechanics and Mechanobiology of Ministry of Education, Beihang University</institution><addr-line>Beijing</addr-line><country>China</country></aff><aff id="aff3"><institution>Key Laboratory of Innovation and Transformation of Advanced Medical Devices, Ministry of Industry and Information Technology</institution><addr-line>Beijing</addr-line><country>China</country></aff><aff id="aff4"><institution>National Medical Innovation Platform for Industry-Education Integration in Advanced Medical Devices (Interdiscipline of Medicine and Engineering)</institution><addr-line>Beijing</addr-line><country>China</country></aff><contrib-group><contrib contrib-type="editor"><name name-style="western"><surname>Schmid</surname><given-names>Maurizio</given-names></name></contrib></contrib-group><contrib-group><contrib contrib-type="reviewer"><name name-style="western"><surname>Yang</surname><given-names>Qi</given-names></name></contrib><contrib contrib-type="reviewer"><name name-style="western"><surname>Xu</surname><given-names>Tong Bill</given-names></name></contrib></contrib-group><author-notes><corresp>Correspondence to Xiaoyu Liu, PhD, School of Biological Science and Medical Engineering, Beihang University, 5 Building in Beihang University, 37 Xueyuan Road, Beijing, China, 86 1082339861; <email>x.y.liu@buaa.edu.cn</email></corresp><fn fn-type="equal" id="equal-contrib1"><label>*</label><p>these authors contributed equally</p></fn></author-notes><pub-date pub-type="collection"><year>2026</year></pub-date><pub-date pub-type="epub"><day>2</day><month>4</month><year>2026</year></pub-date><volume>14</volume><elocation-id>e83128</elocation-id><history><date date-type="received"><day>30</day><month>08</month><year>2025</year></date><date date-type="rev-recd"><day>03</day><month>02</month><year>2026</year></date><date date-type="accepted"><day>15</day><month>02</month><year>2026</year></date></history><copyright-statement>&#x00A9; Xiaofeng Qiao, Shan Tian, Min Tang, Shipei He, Jinghui Wang, Linyuan Fan, Yuanjie Zhu, Zhiyang Zhang, Songjun Du, Chaojie Dong, Yepu Chen, Xiaoyu Liu. Originally published in JMIR Serious Games (<ext-link ext-link-type="uri" xlink:href="https://games.jmir.org">https://games.jmir.org</ext-link>), 2.4.2026. </copyright-statement><copyright-year>2026</copyright-year><license license-type="open-access" xlink:href="https://creativecommons.org/licenses/by/4.0/"><p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (<ext-link ext-link-type="uri" xlink:href="https://creativecommons.org/licenses/by/4.0/">https://creativecommons.org/licenses/by/4.0/</ext-link>), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Serious Games, is properly cited. The complete bibliographic information, a link to the original publication on <ext-link ext-link-type="uri" xlink:href="https://games.jmir.org">https://games.jmir.org</ext-link>, as well as this copyright and license information must be included.</p></license><self-uri xlink:type="simple" xlink:href="https://games.jmir.org/2026/1/e83128"/><abstract><sec><title>Background</title><p>With rapid urbanization, the proliferation of densely arranged buildings and increasingly homogeneous architectural designs has made disorientation and navigation difficulties more common, especially for older adults. Meanwhile, advances in virtual reality technology now allow researchers to create highly immersive navigation games, offering opportunities for assessing cognitive abilities and examining how environmental factors shape navigation behavior.</p></sec><sec><title>Objective</title><p>This study aimed to design a virtual reality&#x2013;based navigation game capable of assessing cognitive abilities through navigation behavior and quantitatively examining how environmental configurations influence navigation patterns in different age groups.</p></sec><sec sec-type="methods"><title>Methods</title><p>We designed a virtual goal&#x2013;directed navigation game and recruited 2 groups, younger adults (n=18) and older adults (n=21), to complete identical wayfinding tasks. Before the formal experiment, participants completed cognitive assessments and received training. To characterize navigational behavior, k-means clustering was applied to classify navigation states and extract behaviorally meaningful navigation measurements, which were then examined for correlations with cognitive test scores. To quantify the effects of environmental structure, space syntax analysis was conducted to calculate line-based and grid-based experienced metrics for each participant, and their associations with navigation performance were examined. Additionally, between-group differences in navigation performance and experienced metrics were evaluated across age groups.</p></sec><sec sec-type="results"><title>Results</title><p>Our results revealed that navigation behavior performance, particularly navigation efficiency, was significantly influenced by cognitive abilities and was strongly associated with several cognitive tests: the Montreal Cognitive Assessment (<italic>r</italic>=0.495, <italic>P</italic>=.04), Trail Making Test Part A (<italic>r</italic>=&#x2212;0.761, <italic>P</italic>=.001), and the Mental Rotation Test (<italic>r</italic>=0.848, <italic>P</italic>&#x003C;.001). In terms of environmental influences, experienced axial integration (EAI) and experienced visual integration (EVI) demonstrated significant age-related differences: EAI (<italic>z</italic>=&#x2013;2.43, <italic>P</italic>=.01) and EVI (<italic>t</italic>=2.48, <italic>P</italic>=.02). Moreover, navigation efficiency exhibited distinct age-specific correlations with experienced metrics: among older adults, navigation efficiency was negatively associated with EVI (<italic>r</italic>=&#x2013;0.48, <italic>P</italic>=.04), and young adults showed negative correlation between navigation efficiency and EAI (<italic>r</italic>=&#x2013;0.64, <italic>P</italic>=.005).</p></sec><sec sec-type="conclusions"><title>Conclusions</title><p>Our findings demonstrate that k-means clustering provides an effective approach for classifying navigation states and extracting quantitative behavioral indicators for assessing cognitive abilities. In addition, the environment-based experienced metrics derived from space syntax analysis revealed distinct age-related navigation patterns, highlighting how spatial configuration shapes wayfinding behavior across age groups. These results establish an important foundation for future applications in clinical cognitive assessment and rehabilitation, as well as the design of age-friendly urban environments.</p></sec></abstract><kwd-group><kwd>spatial navigation</kwd><kwd>environmental structure</kwd><kwd>serious games</kwd><kwd>space syntax</kwd><kwd>cognitive assessment</kwd><kwd>urban planning and design</kwd><kwd>aging</kwd></kwd-group></article-meta></front><body><sec id="s1" sec-type="intro"><title>Introduction</title><p>With rapid urbanization, the proliferation of densely packed buildings with homogeneous architectural designs has increased the likelihood of disorientation and navigation difficulties. Irregular street networks and the frequent modification of road layouts further disrupt the formation of cognitive maps and hinder spatial orientation [<xref ref-type="bibr" rid="ref1">1</xref>]. Spatial navigation, the ability to determine an appropriate route and purposefully move toward a specific destination, has increasingly become a considerable challenge, particularly for older adults [<xref ref-type="bibr" rid="ref2">2</xref>,<xref ref-type="bibr" rid="ref3">3</xref>].</p><p>Significant declines in navigation ability have been recognized as one of the earliest indicators of mild cognitive impairment and Alzheimer disease [<xref ref-type="bibr" rid="ref4">4</xref>,<xref ref-type="bibr" rid="ref5">5</xref>]. Individuals experiencing navigation difficulties tend to confine themselves to familiar environments, resulting in critical limitations on their ability to engage in daily activities [<xref ref-type="bibr" rid="ref5">5</xref>]. Neuroscientific evidence underscores the vulnerability of the entorhinal cortex, a crucial region involved in spatial navigation, to neurodegenerative processes [<xref ref-type="bibr" rid="ref6">6</xref>]. Grid cells within the entorhinal cortex, which are essential for spatial representation and path integration, exhibit notable functional impairments as individuals age. These age-related deficits in grid cell activity may disrupt the formation of stable cognitive maps, leading to difficulties in orientation, route planning, and wayfinding [<xref ref-type="bibr" rid="ref7">7</xref>]. Early and rapid identification of navigation impairments is of critical importance.</p><p>Evidence from previous studies has indicated that behavioral performance in navigation tasks is influenced by both extrinsic factors (eg, spatial properties of the environment, visual accessibility of the landscape) [<xref ref-type="bibr" rid="ref8">8</xref>,<xref ref-type="bibr" rid="ref9">9</xref>] and intrinsic factors (eg, age, cognitive abilities) [<xref ref-type="bibr" rid="ref10">10</xref>,<xref ref-type="bibr" rid="ref11">11</xref>]. These factors jointly determine how individuals gather and interpret spatial information, select navigation strategies, and execute wayfinding decisions. Extensive research has demonstrated that environmental cues [<xref ref-type="bibr" rid="ref12">12</xref>], such as the configuration of landmarks [<xref ref-type="bibr" rid="ref8">8</xref>], locations [<xref ref-type="bibr" rid="ref13">13</xref>], and paths [<xref ref-type="bibr" rid="ref14">14</xref>,<xref ref-type="bibr" rid="ref15">15</xref>], play a critical role in guiding navigation behavior as well as the neural processes underlying spatial orientation [<xref ref-type="bibr" rid="ref16">16</xref>,<xref ref-type="bibr" rid="ref17">17</xref>]. Importantly, age-related differences in the use of environmental cues are well documented [<xref ref-type="bibr" rid="ref18">18</xref>]. Older adults tend to rely more on associating specific directional responses with particular landmark stimuli [<xref ref-type="bibr" rid="ref19">19</xref>]; when geometric cues conflict with landmark cues, they are more likely to maintain the landmark-response association [<xref ref-type="bibr" rid="ref20">20</xref>]. This reliance reflects an egocentric navigation strategy, in which spatial relations are encoded from a body-centered perspective, and actions are guided by subject-to-object associations [<xref ref-type="bibr" rid="ref18">18</xref>,<xref ref-type="bibr" rid="ref21">21</xref>,<xref ref-type="bibr" rid="ref22">22</xref>]. In contrast, younger adults more effectively integrate both geometric and landmark cues, showing greater flexibility in switching between allocentric (map-based) and egocentric (route-based) frames of reference [<xref ref-type="bibr" rid="ref18">18</xref>].</p><p>Although many studies have explored how modifying environmental elements affects navigation behavior, a consistent and intuitive way to characterize spatial properties across different settings is lacking. Space syntax, a well-established framework for quantifying spatial configuration [<xref ref-type="bibr" rid="ref23">23</xref>], offers a powerful solution. It has been widely applied in urban design [<xref ref-type="bibr" rid="ref24">24</xref>], path planning [<xref ref-type="bibr" rid="ref25">25</xref>], and wayfinding research [<xref ref-type="bibr" rid="ref26">26</xref>]. Within this framework, axial map analysis (AMA) [<xref ref-type="bibr" rid="ref13">13</xref>] characterizes spatial structure through line-based representations, whereas visibility graph analysis (VGA) [<xref ref-type="bibr" rid="ref9">9</xref>,<xref ref-type="bibr" rid="ref27">27</xref>,<xref ref-type="bibr" rid="ref28">28</xref>] captures grid-based visibility relationships. Key parameters, including integration, depth, and connectivity, reflect how accessible, shallow, or interconnected different spatial segments are. Prior research has demonstrated the value of these metrics for predicting exploration and wayfinding behavior. Among them, integration is particularly important, as it indicates how accessible a location can be reached from all others and is considered a key determinant of navigation difficulty [<xref ref-type="bibr" rid="ref29">29</xref>]. Moreover, experienced integration has been shown to indicate the extent to which individuals concentrate their movement within globally integrated regions [<xref ref-type="bibr" rid="ref13">13</xref>]. Spending more time in highly integrated areas enables individuals to form more accurate cognitive maps [<xref ref-type="bibr" rid="ref14">14</xref>], resulting in higher experienced integration. Building on this theoretical foundation, we further introduce a series of experienced metrics derived from both axial and visual analyses to quantify the environmental structures that influence participants&#x2019; actual navigation behavior.</p><p>Recent advancements in virtual reality (VR) technology have enabled the creation of highly immersive virtual environments (VEs) that closely replicate the real world with high fidelity [<xref ref-type="bibr" rid="ref30">30</xref>,<xref ref-type="bibr" rid="ref31">31</xref>] and have been widely applied in the field of spatial navigation. Researchers can precisely control environmental variables such as landmark distribution, path complexity, spatial layout, and visual cues in immersive VEs [<xref ref-type="bibr" rid="ref32">32</xref>], making it possible to investigate navigation behavior under the influence of different factors. Moreover, VR technology allows for the collection of a wide range of behavioral data, such as completion time, movement trajectories, error rates, and specific behavioral reactions at given moments [<xref ref-type="bibr" rid="ref33">33</xref>]. By constructing immersive VEs and a task framework, researchers have developed a variety of experimental paradigms designed for specific navigation tasks, including the Virtual Floor Maze Test [<xref ref-type="bibr" rid="ref34">34</xref>], Morris Water Maze task [<xref ref-type="bibr" rid="ref35">35</xref>], and object-location memory task [<xref ref-type="bibr" rid="ref7">7</xref>,<xref ref-type="bibr" rid="ref36">36</xref>]. Efficient VR-based navigation tasks are regarded as valuable tools for assessing and enhancing spatial navigation [<xref ref-type="bibr" rid="ref34">34</xref>].</p><p>In recent years, navigation tasks have increasingly been gamified [<xref ref-type="bibr" rid="ref37">37</xref>,<xref ref-type="bibr" rid="ref38">38</xref>], with successful applications reported in both the health care domain and geographic science research. For example, the mobile game Sea Hero Quest has been widely used for large-scale assessment of human navigation ability [<xref ref-type="bibr" rid="ref39">39</xref>], providing detailed analyses of navigation performance across age groups and cultural backgrounds [<xref ref-type="bibr" rid="ref12">12</xref>,<xref ref-type="bibr" rid="ref40">40</xref>-<xref ref-type="bibr" rid="ref42">42</xref>]. In parallel, gamification approaches combined with location-based services have demonstrated effectiveness in improving users&#x2019; understanding of urban maps, orientation, and wayfinding skills in real-world city environments [<xref ref-type="bibr" rid="ref43">43</xref>,<xref ref-type="bibr" rid="ref44">44</xref>]. The introduction of serious games (SGs) has enhanced the engagement of navigation tasks while preserving scientific validity. By incorporating clearly defined goals, standardized task structures, and meaningful feedback within immersive environments, SG-based VR navigation paradigms provide effective platforms for assessment, training, and rehabilitation [<xref ref-type="bibr" rid="ref45">45</xref>].</p><p>In this study, we intentionally integrated structured navigation tasks into a goal-oriented game framework. We developed a goal-oriented VR-based navigation game, aimed at effectively assessing individual cognitive abilities and exploring how environmental configurations influence navigation behavior. Specifically, the objectives were to (1) evaluate the feasibility and reliability of the VR navigation game as a cognitive assessment tool; (2) compare navigation behaviors between younger and older adults; and (3) examine how key environmental factors shape navigation performance across age groups. These findings are expected to provide methodological support for cognitive assessment and training, as well as empirical evidence to inform spatial planning in aging cities.</p><p>We hypothesized that navigation behavior would be closely associated with cognitive abilities, such that individuals with higher cognitive capacity would demonstrate more efficient wayfinding performance. We further expected that younger and older adults would exhibit distinct navigation patterns, reflecting age-related differences in cognitive processes and strategy use. Finally, we anticipated that quantitatively measured environmental configuration would exert effects on navigation efficiency, and that these effects would differ significantly between age groups.</p></sec><sec id="s2" sec-type="methods"><title>Methods</title><sec id="s2-1"><title>Ethical Considerations</title><p>This research involved human participants. Approval of all ethical and experimental procedures and protocols was granted by the Ethics Committee of Beihang University (BM20230063). Strict privacy protection measures were applied, and all data were anonymized with no personally identifiable information included. Prior to participation, all individuals were fully informed about the study procedures, potential risks, and their right to withdraw at any time without penalty. All participants provided written informed consent and were compensated at a rate of 100 CNY (approximately US $14) per hour.</p><p>Considering that older adults constitute a potentially vulnerable population, additional safeguards were implemented for the older adult group. These included the presence of at least one qualified health professional to provide immediate medical support, repeated verbal reports to monitor fatigue, dizziness, and overall comfort, and the option for participants to terminate the experiment immediately upon experiencing any discomfort.</p></sec><sec id="s2-2"><title>Participants</title><p>The study was conducted with 21 healthy older adults (14 males and 7 females) aged 65 to 80 years (mean age 71.79&#x202F;, SD 4.57 years) and 18 healthy younger adults (10 males and 8 females) aged 21 to 33 years (mean age 26.50, SD&#x202F;2.99 years). All participants had no history of neurological or psychiatric disorders. Older participants were recruited from the community and were at least 65 years old and able to complete daily activities independently. One older participant with a history of cerebral hemorrhage was excluded prior to the study, and another withdrew due to intolerance to VR-induced dizziness. Younger participants were recruited through public advertisements posted on the university campus. Individuals who expressed interest were screened using the same inclusion criteria applied to all participants, including normal or corrected-to-normal vision, no history of neurological or psychiatric disorders, and no prior diagnosis of cognitive impairment. Additionally, individual information, encompassing age, gender, education, and prior VR experience, was systematically gathered from each participant.</p></sec><sec id="s2-3"><title>System Configuration</title><p>The virtual navigation game used in this study was developed with Unity 2019.4 (Unity Technologies) and deployed via the HTC VIVE Pro head-mounted display. The visual scene was rendered at 90 Hz and presented at a single-eye resolution of 1440&#x00D7;1600 pixels (2880&#x00D7;1600 binocular) with a 110&#x00B0; field of view, providing an immersive experience. For the experimental setup, the head-mounted display was connected to a high-performance desktop PC via a wired connection using the standard VIVE link box to ensure low latency and stable rendering. The system was equipped with an Intel Core i7-9700K CPU, an NVIDIA GeForce RTX 2070 GPU, 32 GB of RAM, and the Windows 10 operating system, enabling smooth rendering and reliable behavioral data acquisition.</p><p>The VE was designed as a simplified yet ecologically valid urban block (200 vm&#x00D7;150 vm, virtual meters), featuring residential buildings, street networks, greenery, and small parks. The intentional integration of these elements provided sufficient spatial complexity to elicit meaningful wayfinding behavior while minimizing potential discomfort or anxiety associated with navigation in a VE [<xref ref-type="bibr" rid="ref46">46</xref>].</p><p>Given that older participants may experience limitations in sustained physical walking, we adopted a hybrid navigation control scheme combining the handheld controller and natural facing direction [<xref ref-type="bibr" rid="ref47">47</xref>]. The VIVE controller features an ergonomic design with a circular touchpad. To minimize operational complexity, participants were instructed to press the touchpad to move forward in the direction they were facing, and releasing it stopped movement immediately. Horizontal body rotation was mapped one-to-one to virtual orientation, allowing participants to turn naturally without additional button input. Throughout the experiment, all participants were instructed to avoid movement in the physical space and to navigate using the controller and body rotation only. Most participants completed the game while standing, whereas three older participants, due to advanced age, performed the experiment while seated in a swivel chair that allowed 360&#x00B0; physical rotation. This simplified control scheme reduced cognitive and motor demands while ensuring consistent interaction across participants.</p></sec><sec id="s2-4"><title>Experimental Procedure</title><p>The experiment consisted of three phases: (1) neuropsychological assessment, (2) pretraining, and (3) formal experiment. The neuropsychological assessment phase aimed to evaluate the cognitive abilities of older participants using standardized measures of general cognitive skills and empirically validated tests. During the pretraining phase, participants were instructed to navigate freely within the VE for 15 minutes. Navigation proficiency was considered achieved when participants demonstrated the ability to independently and fluently operate the controller and navigate to designated target locations, qualifying them to proceed to the formal experimental phase. In the formal experimental phase, participants were required to sequentially navigate to multiple target locations, with spatiotemporal data recorded at fixed intervals throughout the navigation process. Older participants completed all 3 phases, while younger participants, who were healthy university students with presumed normal cognitive function, only participated in phases 2 and 3.</p><p>Before the formal navigation tasks, general cognitive functioning was quantified using the Mini-Mental State Examination (MMSE) [<xref ref-type="bibr" rid="ref48">48</xref>,<xref ref-type="bibr" rid="ref49">49</xref>] and the Montreal Cognitive Assessment (MoCA) [<xref ref-type="bibr" rid="ref50">50</xref>]. Specifically, the MMSE administered in this study was the standardized Chinese adaptation based on the international MMSE framework, which has been validated through prior pilot testing and is widely used in studies involving Chinese populations [<xref ref-type="bibr" rid="ref48">48</xref>,<xref ref-type="bibr" rid="ref49">49</xref>,<xref ref-type="bibr" rid="ref51">51</xref>]. The MoCA Full Chinese (Beijing) Version 7.1 was the official Beijing version published on the MoCA website [<xref ref-type="bibr" rid="ref52">52</xref>]. MoCA scores were adjusted according to participants&#x2019; education level, with one additional point added for individuals with 12 years of formal education or fewer. Nonverbal memory was further assessed using the delayed recall condition of the Rey-Osterrieth Complex Figure Test, while visuospatial functioning was evaluated through the Copy condition of the Rey-Osterrieth Complex Figure (ROCF-C) test [<xref ref-type="bibr" rid="ref53">53</xref>]. To examine executive function and working memory, Part A of the Trail Making Test (TMTA) was administered [<xref ref-type="bibr" rid="ref54">54</xref>]. Additionally, spatial rotation ability was assessed using a standardized Mental Rotation Test (MRT) [<xref ref-type="bibr" rid="ref55">55</xref>]. These assessments provided a comprehensive cognitive profile for each participant prior to engagement in the virtual navigation experiment.</p><p>During the pre-experimental phase, identically shaped but differently colored target objects were placed at various locations within the VE (<xref ref-type="fig" rid="figure1">Figure 1</xref>). Participants were instructed to memorize the spatial locations corresponding to each colored object. The specific verbal instructions can be found in <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref>. During the formal experimental phase, 5 target locations previously encountered during the training phase were used (<xref ref-type="fig" rid="figure2">Figure 2C</xref>). Target locations were presented in a fixed sequential order, allowing for consistent route comparisons across participants. Each trial began with a fixation cross (1000 ms), followed by the target cue (2000 ms), another fixation cross (1000 ms), and then participants were required to navigate to the designated target location (<xref ref-type="fig" rid="figure2">Figure 2A</xref>). Upon reaching the target, the current trial was concluded, and the next trial was automatically initiated. This sequence continued until all tasks were completed. Participants were instructed to navigate as quickly and accurately as possible (<xref ref-type="fig" rid="figure2">Figure 2B</xref>), and a congratulatory animation was displayed after each trial to prevent boredom.</p><fig position="float" id="figure1"><label>Figure 1.</label><caption><p>Pretraining phase: top-down view of the scene (middle) and views of the target from four different locations (left and right).</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="games_v14i1e83128_fig01.png"/></fig><fig position="float" id="figure2"><label>Figure 2.</label><caption><p>Formal experiment: (A) The sequence of a single trial. Following the presentation of a cue, participants were required to navigate to the target location. (B) Photograph of an older participant during the experiment. (C) Layout of the virtual city environment, including the shortest route planned to all target locations.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="games_v14i1e83128_fig02.png"/></fig></sec><sec id="s2-5"><title>Environmental Structure Analysis</title><sec id="s2-5-1"><title>Space Syntax</title><p>Space Syntax provides a robust and systematic framework for examining the spatial arrangement of environments and the interrelationships between spaces. This methodology has been extensively applied to explore how spatial configurations affect human behavior [<xref ref-type="bibr" rid="ref26">26</xref>,<xref ref-type="bibr" rid="ref56">56</xref>,<xref ref-type="bibr" rid="ref57">57</xref>], particularly how individuals encode, represent, and retrieve spatial information during navigation. In this study, both AMA and the VGA of the VE were carried out using the DepthmapX software (Space Syntax Laboratory).</p></sec><sec id="s2-5-2"><title>Axial Map Analysis</title><p>The axial map provides a graphical representation of spatial structure, where streets within an urban environment are abstracted as axial lines, defined as the minimal set of longest straight lines of sight from a given location [<xref ref-type="bibr" rid="ref23">23</xref>,<xref ref-type="bibr" rid="ref58">58</xref>]. These lines collectively cover the entire space, capturing all possible movement paths. Based on the 2D layout of the VE street network, we constructed an axial map with a total of 12 axial lines delineated (<xref ref-type="fig" rid="figure3">Figure 3</xref>). For analytical purposes, key metrics such as axial integration, axial connectivity, and axial mean depth were calculated to characterize the accessibility and flow potential of individual environmental segments.</p><fig position="float" id="figure3"><label>Figure 3.</label><caption><p>The axial map constructed from the 2D layout of the virtual environment street network.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="games_v14i1e83128_fig03.png"/></fig></sec><sec id="s2-5-3"><title>Visibility Graph Analysis</title><p>VGA provides an analytical approach for quantifying the spatial characteristics of an environment by assessing the extent to which any given point within the space is visible from any other point [<xref ref-type="bibr" rid="ref59">59</xref>]. In this study, the structural layout of the VE was first modeled using AutoCAD. Subsequently, a 1 m&#x00D7;1 &#x202F;m grid was generated in DepthmapX to perform the analysis. This process enables the computation of key spatial metrics at each node, including visual integration, visual connectivity, and visual mean depth.</p></sec></sec><sec id="s2-6"><title>Measurement</title><sec id="s2-6-1"><title>Metrics Setting</title><p>A set of behavioral and environment-based metrics was calculated in the virtual navigation game to evaluate the effectiveness of the game in assessing cognitive abilities and to examine age-related differences in navigation performance and sensitivity to environmental features. Spatiotemporal data were recorded for each participant at a fixed sampling frequency. Based on these data, the navigation behavioral metrics and environment-based navigation metrics were defined as follows.</p></sec><sec id="s2-6-2"><title>Wayfinding, Transition, and Moving States</title><p>During navigation, we continuously recorded each participant&#x2019;s spatial position <inline-formula><mml:math id="ieqn1"><mml:mstyle><mml:mrow><mml:mstyle displaystyle="false"><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mi>x</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mi>y</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mstyle></mml:mrow></mml:mstyle></mml:math></inline-formula> at timestamp <inline-formula><mml:math id="ieqn2"><mml:msub><mml:mrow><mml:mi>t</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> . For each recorded point <inline-formula><mml:math id="ieqn3"><mml:mi>i</mml:mi></mml:math></inline-formula> (excluding the starting and ending points), the instantaneous speed was computed as the average of the velocity estimates derived from two adjacent time intervals.</p><disp-formula id="equWL1"><mml:math id="eqn1"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:mstyle displaystyle="true" scriptlevel="0"><mml:msub><mml:mi>v</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mfrac><mml:msqrt><mml:msup><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mi>x</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>&#x2212;</mml:mo><mml:msub><mml:mi>x</mml:mi><mml:mrow><mml:mi>i</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup><mml:mo>+</mml:mo><mml:msup><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mi>y</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>&#x2212;</mml:mo><mml:msub><mml:mi>y</mml:mi><mml:mrow><mml:mi>i</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup></mml:msqrt><mml:mrow><mml:msub><mml:mi>t</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>&#x2212;</mml:mo><mml:msub><mml:mi>t</mml:mi><mml:mrow><mml:mi>i</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:mfrac><mml:mo>+</mml:mo><mml:mfrac><mml:msqrt><mml:msup><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mi>x</mml:mi><mml:mrow><mml:mi>i</mml:mi><mml:mo>+</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>&#x2212;</mml:mo><mml:msub><mml:mi>x</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup><mml:mo>+</mml:mo><mml:msup><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mi>y</mml:mi><mml:mrow><mml:mi>i</mml:mi><mml:mo>+</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>&#x2212;</mml:mo><mml:msub><mml:mi>y</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup></mml:msqrt><mml:mrow><mml:msub><mml:mi>t</mml:mi><mml:mrow><mml:mi>i</mml:mi><mml:mo>+</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>&#x2212;</mml:mo><mml:msub><mml:mi>t</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mfrac></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mrow><mml:mo>/</mml:mo></mml:mrow><mml:mn>2</mml:mn></mml:mstyle></mml:mrow></mml:mstyle></mml:math></disp-formula><p>The instantaneous speed series for each participant was then classified into 3 distinct categories using the k-means clustering algorithm, yielding low, medium, and high speed clusters. These clusters corresponded to meaningful navigation states: Wayfinding (low speed: searching, hesitation, or reorientation), Transition (medium speed: directional adjustment or preparatory acceleration), and Moving (high speed: continuous, goal-directed locomotion). This classification allowed us to segment the entire trajectory into functionally interpretable behavioral states.</p></sec><sec id="s2-6-3"><title>Navigation Efficiency</title><p>Navigation efficiency is defined as the ratio of effective navigation time (moving time and transition time) to the total duration of the navigation game (navigation time).</p><disp-formula id="equWL2"><mml:math id="eqn2"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mi>N</mml:mi><mml:mi>a</mml:mi><mml:mi>v</mml:mi><mml:mi>i</mml:mi><mml:mi>g</mml:mi><mml:mi>a</mml:mi><mml:mi>t</mml:mi><mml:mi>i</mml:mi><mml:mi>o</mml:mi><mml:mi>n</mml:mi><mml:mtext>&#x00A0;</mml:mtext><mml:mi>e</mml:mi><mml:mi>f</mml:mi><mml:mi>f</mml:mi><mml:mi>i</mml:mi><mml:mi>c</mml:mi><mml:mi>i</mml:mi><mml:mi>e</mml:mi><mml:mi>n</mml:mi><mml:mi>c</mml:mi><mml:mi>y</mml:mi><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mi>M</mml:mi><mml:mi>o</mml:mi><mml:mi>v</mml:mi><mml:mi>i</mml:mi><mml:mi>n</mml:mi><mml:mi>g</mml:mi><mml:mtext>&#x00A0;</mml:mtext><mml:mi>t</mml:mi><mml:mi>i</mml:mi><mml:mi>m</mml:mi><mml:mi>e</mml:mi><mml:mo>+</mml:mo><mml:mi>T</mml:mi><mml:mi>r</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi><mml:mi>s</mml:mi><mml:mi>i</mml:mi><mml:mi>t</mml:mi><mml:mi>i</mml:mi><mml:mi>o</mml:mi><mml:mi>n</mml:mi><mml:mtext>&#x00A0;</mml:mtext><mml:mi>t</mml:mi><mml:mi>i</mml:mi><mml:mi>m</mml:mi><mml:mi>e</mml:mi></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mrow><mml:mi>N</mml:mi><mml:mi>a</mml:mi><mml:mi>v</mml:mi><mml:mi>i</mml:mi><mml:mi>g</mml:mi><mml:mi>a</mml:mi><mml:mi>t</mml:mi><mml:mi>i</mml:mi><mml:mi>o</mml:mi><mml:mi>n</mml:mi><mml:mtext>&#x00A0;</mml:mtext><mml:mi>t</mml:mi><mml:mi>i</mml:mi><mml:mi>m</mml:mi><mml:mi>e</mml:mi></mml:mrow></mml:mfrac></mml:mstyle></mml:mrow></mml:mstyle></mml:math></disp-formula><p>This metric serves as an indicator of the capability to navigate effectively within the environment. Higher values reflect a more streamlined and goal-oriented navigation process, characterized by reduced time spent on behaviors such as hesitation, reorientation, or uncertainty. By quantifying navigation efficiency, we can evaluate the extent to which individuals construct accurate cognitive maps and adopt effective spatial strategies during navigation tasks.</p></sec><sec id="s2-6-4"><title>Experienced Axial Metrics</title><p>Axial analysis is derived from a representation of the environment using the fewest and longest lines of sight, which capture key structural characteristics of an urban street network. In this study, we used the axial values computed from the axial map shown in <xref ref-type="fig" rid="figure3">Figure 3</xref>. Three axial measures were considered: axial integration, which quantifies the accessibility of an axial line from the rest of the system; axial connectivity, which denotes the number of axial lines directly linked to a given line; and axial mean depth, which represents the arithmetic mean of the topological depths from each axial line to all others. To examine how participants&#x2019; navigation behavior was shaped by the underlying street configuration, we computed the average value of each metric along the participant&#x2019;s trajectory, resulting in 3 experienced axial metrics: experienced axial integration (EAI), experienced axial connectivity (EAC), and experienced mean axial depth (EMAD).</p><disp-formula id="equWL3"><mml:math id="eqn3"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mi>E</mml:mi><mml:mi>A</mml:mi><mml:mi>I</mml:mi><mml:mo>=</mml:mo><mml:mfrac><mml:mn>1</mml:mn><mml:mi>n</mml:mi></mml:mfrac><mml:munderover><mml:mo>&#x2211;</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>n</mml:mi></mml:mrow></mml:munderover><mml:msub><mml:mi>I</mml:mi><mml:mrow><mml:mi>A</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mi>x</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mi>y</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mstyle></mml:mstyle></mml:mrow></mml:mstyle></mml:math></disp-formula><disp-formula id="equWL4"><mml:math id="eqn4"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mi>E</mml:mi><mml:mi>A</mml:mi><mml:mi>C</mml:mi><mml:mo>=</mml:mo><mml:mfrac><mml:mn>1</mml:mn><mml:mi>n</mml:mi></mml:mfrac><mml:munderover><mml:mo>&#x2211;</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>n</mml:mi></mml:mrow></mml:munderover><mml:msub><mml:mi>C</mml:mi><mml:mrow><mml:mi>A</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mi>x</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mi>y</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mstyle></mml:mstyle></mml:mrow></mml:mstyle></mml:math></disp-formula><disp-formula id="equWL5"><mml:math id="eqn5"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mi>E</mml:mi><mml:mi>M</mml:mi><mml:mi>A</mml:mi><mml:mi>D</mml:mi><mml:mo>=</mml:mo><mml:mfrac><mml:mn>1</mml:mn><mml:mi>n</mml:mi></mml:mfrac><mml:munderover><mml:mo>&#x2211;</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>n</mml:mi></mml:mrow></mml:munderover><mml:msub><mml:mi>D</mml:mi><mml:mrow><mml:mi>A</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mi>x</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mi>y</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mstyle></mml:mstyle></mml:mrow></mml:mstyle></mml:math></disp-formula><p>The point <inline-formula><mml:math id="ieqn4"><mml:mo>(</mml:mo><mml:msub><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mrow><mml:mi>y</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>)</mml:mo></mml:math></inline-formula> denotes the coordinates of the participant at sample <italic>i</italic>, and <inline-formula><mml:math id="ieqn5"><mml:msub><mml:mrow><mml:mi>I</mml:mi></mml:mrow><mml:mrow><mml:mi>A</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, <inline-formula><mml:math id="ieqn6"><mml:msub><mml:mrow><mml:mi>C</mml:mi></mml:mrow><mml:mrow><mml:mi>A</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, and <inline-formula><mml:math id="ieqn7"><mml:msub><mml:mrow><mml:mi>D</mml:mi></mml:mrow><mml:mrow><mml:mi>A</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> represent the integration, connectivity, and mean depth values of the axial line associated with that position. The variable n indicates the total number of sampled points along the participant&#x2019;s trajectory.</p></sec><sec id="s2-6-5"><title>Experienced Visual Metrics</title><p>VGA represents an environment as a grid-based visibility network, enabling the quantification of visual relationships between locations. The visual integration, visual connectivity, and visual mean depth values used in this study were derived from the VGA map shown in <xref ref-type="fig" rid="figure4">Figure 4</xref>. To examine how participants&#x2019; navigation behavior was influenced by visually accessible spatial properties, we computed the average value of each VGA metric along the participant&#x2019;s trajectory, yielding experienced visual integration (EVI), experienced visual connectivity, and experienced visual mean depth.</p><disp-formula id="equWL6"><mml:math id="eqn6"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mi>E</mml:mi><mml:mi>V</mml:mi><mml:mi>I</mml:mi><mml:mo>=</mml:mo><mml:mfrac><mml:mn>1</mml:mn><mml:mi>n</mml:mi></mml:mfrac><mml:munderover><mml:mo>&#x2211;</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>n</mml:mi></mml:mrow></mml:munderover><mml:msub><mml:mi>I</mml:mi><mml:mrow><mml:mi>v</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mi>x</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mi>y</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mstyle></mml:mstyle></mml:mrow></mml:mstyle></mml:math></disp-formula><disp-formula id="equWL7"><mml:math id="eqn7"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mi>E</mml:mi><mml:mi>V</mml:mi><mml:mi>C</mml:mi><mml:mo>=</mml:mo><mml:mfrac><mml:mn>1</mml:mn><mml:mi>n</mml:mi></mml:mfrac><mml:munderover><mml:mo>&#x2211;</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>n</mml:mi></mml:mrow></mml:munderover><mml:msub><mml:mi>C</mml:mi><mml:mrow><mml:mi>v</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mi>x</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mi>y</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mstyle></mml:mstyle></mml:mrow></mml:mstyle></mml:math></disp-formula><disp-formula id="equWL8"><mml:math id="eqn8"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mi>E</mml:mi><mml:mi>V</mml:mi><mml:mi>M</mml:mi><mml:mi>D</mml:mi><mml:mo>=</mml:mo><mml:mfrac><mml:mn>1</mml:mn><mml:mi>n</mml:mi></mml:mfrac><mml:munderover><mml:mo>&#x2211;</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>n</mml:mi></mml:mrow></mml:munderover><mml:msub><mml:mi>D</mml:mi><mml:mrow><mml:mi>v</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mi>x</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mi>y</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mstyle></mml:mstyle></mml:mrow></mml:mstyle></mml:math></disp-formula><p><inline-formula><mml:math id="ieqn8"><mml:msub><mml:mrow><mml:mi>I</mml:mi></mml:mrow><mml:mrow><mml:mi>v</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, <inline-formula><mml:math id="ieqn9"><mml:msub><mml:mrow><mml:mi>C</mml:mi></mml:mrow><mml:mrow><mml:mi>v</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, and <inline-formula><mml:math id="ieqn10"><mml:msub><mml:mrow><mml:mi>D</mml:mi></mml:mrow><mml:mrow><mml:mi>v</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> represent visual integration, visual connectivity, and visual mean depth values associated with participants&#x2019; position <inline-formula><mml:math id="ieqn11"><mml:mo>(</mml:mo><mml:msub><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mrow><mml:mi>y</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>)</mml:mo></mml:math></inline-formula> in the VGA grid.</p><fig position="float" id="figure4"><label>Figure 4.</label><caption><p>The visualization of the virtual urban spatial structure using visibility graph analysis.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="games_v14i1e83128_fig04.png"/></fig></sec></sec><sec id="s2-7"><title>Statistical Analysis</title><p>Data from the experiment were processed using custom scripts developed in MATLAB, and all statistical analyses were performed using SPSS (IBM Corp). The normality of each variable was assessed through visual inspection of histograms and the Shapiro&#x2013;Wilk test. Spearman&#x2019;s rank correlation analyses were conducted to examine the associations between navigation-related behavioral measures and cognitive test scores. To evaluate age-related differences, independent-samples <italic>t</italic> tests were applied to variables that met normality assumptions, whereas the Mann-Whitney <italic>U</italic> test was used for variables that were not normally distributed. In addition, correlations between navigation efficiency and the various experienced spatial metrics were examined using Pearson correlations for normally distributed variables and Spearman rank correlations for nonnormally distributed variables.</p></sec></sec><sec id="s3" sec-type="results"><title>Results</title><sec id="s3-1"><title>Participants</title><p>All 19 older participants and 18 young participants successfully completed the experiment without experiencing severe motion sickness. <xref ref-type="table" rid="table1">Table 1</xref> summarizes the cognitive characteristics of the older group, including the means and SDs for various cognitive assessment scales. Notably, none of the participants had prior experience with VR technology, which reduced potential biases associated with equipment familiarity.</p><p>After clustering the data, points labeled as wayfinding state exhibited clear spatial concentration and age-related differences (<xref ref-type="fig" rid="figure5">Figure 5</xref>). This spatial density map visualizes wayfinding patterns, offering an intuitive overview of participants&#x2019; spatial behavior based on spatiotemporal features. The red-outlined areas indicate the top 1% of regions with the highest density values, highlighting zones of intensified wayfinding activity within the VE.</p><table-wrap id="t1" position="float"><label>Table 1.</label><caption><p>Characteristics of participants of different groups.</p></caption><table id="table1" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom"/><td align="left" valign="bottom">Older adults</td><td align="left" valign="bottom">Young adults</td></tr></thead><tbody><tr><td align="left" valign="top">Age (years), mean (SD)</td><td align="left" valign="top">71.79 (4.57)</td><td align="left" valign="top">26.50&#x202F;(2.99)</td></tr><tr><td align="left" valign="top">Sex (male/female), n</td><td align="left" valign="top">14/5</td><td align="left" valign="top">10/8</td></tr><tr><td align="left" valign="top">Education &#x003E;12 years (yes/no), n</td><td align="left" valign="top">11/8</td><td align="left" valign="top">18/0</td></tr><tr><td align="left" valign="top">VR<sup><xref ref-type="table-fn" rid="table1fn1">a</xref></sup> experience (yes/no), n</td><td align="left" valign="top">0/19</td><td align="left" valign="top">0/18</td></tr><tr><td align="left" valign="top">MoCA<sup><xref ref-type="table-fn" rid="table1fn2">b</xref></sup>, mean (SD)</td><td align="left" valign="top">21.84 (5.47)</td><td align="left" valign="top">&#x2014;<sup><xref ref-type="table-fn" rid="table1fn3">c</xref></sup></td></tr><tr><td align="left" valign="top">MMSE<sup><xref ref-type="table-fn" rid="table1fn4">d</xref></sup>, mean (SD)</td><td align="left" valign="top">27.16 (3.60)</td><td align="left" valign="top">&#x2014;</td></tr><tr><td align="left" valign="top">TMTA<sup><xref ref-type="table-fn" rid="table1fn5">e</xref></sup>, mean (SD)</td><td align="left" valign="top">62.06 (21.27)</td><td align="left" valign="top">&#x2014;</td></tr><tr><td align="left" valign="top">ROCF-D<sup><xref ref-type="table-fn" rid="table1fn6">f</xref></sup>/ROCF-C<sup><xref ref-type="table-fn" rid="table1fn7">g</xref></sup>, mean (SD)</td><td align="left" valign="top">31.63 (4.59)/18.60 (5.62)</td><td align="left" valign="top">&#x2014;</td></tr><tr><td align="left" valign="top">MRT<sup><xref ref-type="table-fn" rid="table1fn8">h</xref></sup>, mean (SD)</td><td align="left" valign="top">68.58 (13.62)</td><td align="left" valign="top">&#x2014;</td></tr></tbody></table><table-wrap-foot><fn id="table1fn1"><p><sup>a</sup>VR: virtual reality.</p></fn><fn id="table1fn2"><p><sup>b</sup>MoCA: Montreal Cognitive Assessment.</p></fn><fn id="table1fn3"><p><sup>c</sup>Not available.</p></fn><fn id="table1fn4"><p><sup>d</sup>MMSE: Mini-Mental State Examination.</p></fn><fn id="table1fn5"><p><sup>e</sup>TMTA: Part A of the Trail Making Test.</p></fn><fn id="table1fn6"><p><sup>f</sup>ROCF-D: delayed recall condition of the Rey-Osterrieth Complex Figure.</p></fn><fn id="table1fn7"><p><sup>g</sup>ROCF-C: Copy condition of the Rey-Osterrieth Complex Figure.</p></fn><fn id="table1fn8"><p><sup>h</sup>MRT: Mental Rotation Test.</p></fn></table-wrap-foot></table-wrap><fig position="float" id="figure5"><label>Figure 5.</label><caption><p>Spatial density distribution of wayfinding points (left: older adults; right: young adults).</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="games_v14i1e83128_fig05.png"/></fig></sec><sec id="s3-2"><title>The Correlation Between Navigation Behavior and Cognitive Abilities in Older Participants</title><p>Spearman rank correlation analysis was conducted to examine the relationships between navigation behavior metrics and cognitive assessment scores (<xref ref-type="table" rid="table2">Table 2</xref>). No significant correlations were observed between either distance or error and any of the cognitive scales. However, temporal-based navigation parameters performed well. Navigation time was significantly associated with both TMTA (<italic>r</italic>=0.503, <italic>P</italic>=.047) and the MRT (<italic>r</italic>=&#x2212;0.526, <italic>P</italic>=.03). Wayfinding time showed strong correlations with several cognitive measures, including TMTA (<italic>r</italic>=0.716, <italic>P</italic>=.002), MoCA (<italic>r</italic>=&#x2212;0.549, <italic>P</italic>=.02), ROCF-C (<italic>r</italic>=&#x2212;0.588, <italic>P</italic>=.01), MMSE (<italic>r</italic>=&#x2212;0.458, <italic>P</italic>=.046) and MRT (<italic>r</italic>=&#x2212;0.751, <italic>P</italic>&#x003C;.001). Additionally, navigation efficiency demonstrated significant associations with MoCA (<italic>r</italic>=0.495, <italic>P</italic>=.04) and ROCF-C (<italic>r</italic>=0.658, <italic>P</italic>=.003) scores, and particularly strong correlations with TMTA (<italic>r</italic>=&#x2212;0.761, <italic>P</italic>=.001) and MRT (<italic>r</italic>=0.848, <italic>P</italic>&#x003C;.001). These findings confirmed that cognitive abilities play a crucial role in influencing navigation efficiency, including navigation time and wayfinding time.</p><table-wrap id="t2" position="float"><label>Table 2.</label><caption><p>The correlation between spatial navigation and cognitive scales.</p></caption><table id="table2" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom"/><td align="left" valign="bottom">MoCA<sup><xref ref-type="table-fn" rid="table2fn1">a</xref></sup></td><td align="left" valign="bottom">MMSE<sup><xref ref-type="table-fn" rid="table2fn2">b</xref></sup></td><td align="left" valign="bottom">ROCF<sup><xref ref-type="table-fn" rid="table2fn3">c</xref></sup>-C</td><td align="left" valign="bottom">ROCF-D<sup><xref ref-type="table-fn" rid="table2fn4">d</xref></sup></td><td align="left" valign="bottom">TMTA<sup><xref ref-type="table-fn" rid="table2fn5">e</xref></sup></td><td align="left" valign="bottom">MRT<sup><xref ref-type="table-fn" rid="table2fn6">f</xref></sup></td><td align="left" valign="bottom">Distance</td><td align="left" valign="bottom">Error</td><td align="left" valign="bottom">Navigation time</td><td align="left" valign="bottom">Wayfinding time</td></tr></thead><tbody><tr><td align="left" valign="top" colspan="11">MMSE</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content><italic>r</italic></td><td align="left" valign="top">0.885<sup><xref ref-type="table-fn" rid="table2fn7">g</xref></sup></td><td align="left" valign="top">1</td><td align="left" valign="top"/><td align="left" valign="top"/><td align="left" valign="top"/><td align="left" valign="top"/><td align="left" valign="top"/><td align="left" valign="top"/><td align="left" valign="top"/><td align="left" valign="top"/></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content><italic>P</italic> value</td><td align="left" valign="top">&#x003C;.001</td><td align="left" valign="top">&#x2014;<sup><xref ref-type="table-fn" rid="table2fn8">h</xref></sup></td><td align="left" valign="top"/><td align="left" valign="top"/><td align="left" valign="top"/><td align="left" valign="top"/><td align="left" valign="top"/><td align="left" valign="top"/><td align="left" valign="top"/><td align="left" valign="top"/></tr><tr><td align="left" valign="top" colspan="11">ROCF-C</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content><italic>r</italic></td><td align="left" valign="top">0.470<sup><xref ref-type="table-fn" rid="table2fn9">i</xref></sup></td><td align="left" valign="top">0.364</td><td align="left" valign="top">1</td><td align="left" valign="top"/><td align="left" valign="top"/><td align="left" valign="top"/><td align="left" valign="top"/><td align="left" valign="top"/><td align="left" valign="top"/><td align="left" valign="top"/></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content><italic>P</italic> value</td><td align="left" valign="top">.049</td><td align="left" valign="top">.14</td><td align="left" valign="top">&#x2014;</td><td align="left" valign="top"/><td align="left" valign="top"/><td align="left" valign="top"/><td align="left" valign="top"/><td align="left" valign="top"/><td align="left" valign="top"/><td align="left" valign="top"/></tr><tr><td align="left" valign="top" colspan="11">ROCF-D</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content><italic>r</italic></td><td align="left" valign="top">0.344</td><td align="left" valign="top">0.084</td><td align="left" valign="top">0.603<sup><xref ref-type="table-fn" rid="table2fn10">j</xref></sup></td><td align="left" valign="top">1</td><td align="left" valign="top"/><td align="left" valign="top"/><td align="left" valign="top"/><td align="left" valign="top"/><td align="left" valign="top"/><td align="left" valign="top"/></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content><italic>P</italic> value</td><td align="left" valign="top">.16</td><td align="left" valign="top">.74</td><td align="left" valign="top">.008</td><td align="left" valign="top">&#x2014;</td><td align="left" valign="top"/><td align="left" valign="top"/><td align="left" valign="top"/><td align="left" valign="top"/><td align="left" valign="top"/><td align="left" valign="top"/></tr><tr><td align="left" valign="top" colspan="11">TMTA</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content><italic>r</italic></td><td align="left" valign="top">&#x2212;0.663<sup><xref ref-type="table-fn" rid="table2fn10">j</xref></sup></td><td align="left" valign="top">&#x2212;0.496</td><td align="left" valign="top">&#x2212;0.380</td><td align="left" valign="top">&#x2212;0.081</td><td align="left" valign="top">1</td><td align="left" valign="top"/><td align="left" valign="top"/><td align="left" valign="top"/><td align="left" valign="top"/><td align="left" valign="top"/></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content><italic>P</italic> value</td><td align="left" valign="top">.005</td><td align="left" valign="top">.05</td><td align="left" valign="top">.15</td><td align="left" valign="top">.76</td><td align="left" valign="top">&#x2014;</td><td align="left" valign="top"/><td align="left" valign="top"/><td align="left" valign="top"/><td align="left" valign="top"/><td align="left" valign="top"/></tr><tr><td align="left" valign="top" colspan="11">MRT</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content><italic>r</italic></td><td align="left" valign="top">0.409</td><td align="left" valign="top">0.301</td><td align="left" valign="top">0.701<sup><xref ref-type="table-fn" rid="table2fn10">j</xref></sup></td><td align="left" valign="top">0.349</td><td align="left" valign="top">&#x2212;0.591<sup><xref ref-type="table-fn" rid="table2fn9">i</xref></sup></td><td align="left" valign="top">1</td><td align="left" valign="top"/><td align="left" valign="top"/><td align="left" valign="top"/><td align="left" valign="top"/></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content><italic>P</italic> value</td><td align="left" valign="top">.09</td><td align="left" valign="top">.22</td><td align="left" valign="top">.001</td><td align="left" valign="top">.16</td><td align="left" valign="top">.02</td><td align="left" valign="top">&#x2014;</td><td align="left" valign="top"/><td align="left" valign="top"/><td align="left" valign="top"/><td align="left" valign="top"/></tr><tr><td align="left" valign="top" colspan="11">Distance</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content><italic>r</italic></td><td align="left" valign="top">&#x2212;0.287</td><td align="left" valign="top">&#x2212;0.201</td><td align="left" valign="top">&#x2212;0.013</td><td align="left" valign="top">&#x2212;0.122</td><td align="left" valign="top">&#x2212;0.105</td><td align="left" valign="top">0.103</td><td align="left" valign="top">1</td><td align="left" valign="top"/><td align="left" valign="top"/><td align="left" valign="top"/></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content><italic>P</italic> value</td><td align="left" valign="top">.25</td><td align="left" valign="top">.42</td><td align="left" valign="top">.96</td><td align="left" valign="top">.63</td><td align="left" valign="top">.70</td><td align="left" valign="top">.68</td><td align="left" valign="top">&#x2014;</td><td align="left" valign="top"/><td align="left" valign="top"/><td align="left" valign="top"/></tr><tr><td align="left" valign="top" colspan="11">Error</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content><italic>r</italic></td><td align="left" valign="top">&#x2212;0.278</td><td align="left" valign="top">&#x2212;0.098</td><td align="left" valign="top">0.049</td><td align="left" valign="top">&#x2212;0.203</td><td align="left" valign="top">&#x2212;0.07</td><td align="left" valign="top">0.313</td><td align="left" valign="top">0.873<sup><xref ref-type="table-fn" rid="table2fn7">g</xref></sup></td><td align="left" valign="top">1</td><td align="left" valign="top"/><td align="left" valign="top"/></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content><italic>P</italic> value</td><td align="left" valign="top">.26</td><td align="left" valign="top">.70</td><td align="left" valign="top">.85</td><td align="left" valign="top">.42</td><td align="left" valign="top">.80</td><td align="left" valign="top">.21</td><td align="left" valign="top">&#x003C;.001</td><td align="left" valign="top">&#x2014;</td><td align="left" valign="top"/><td align="left" valign="top"/></tr><tr><td align="left" valign="top" colspan="11">Navigation time</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content><italic>r</italic></td><td align="left" valign="top">&#x2212;0.399</td><td align="left" valign="top">&#x2212;0.286</td><td align="left" valign="top">&#x2212;0.356</td><td align="left" valign="top">&#x2212;0.247</td><td align="left" valign="top">0.503<sup><xref ref-type="table-fn" rid="table2fn9">i</xref></sup></td><td align="left" valign="top">&#x2212;0.526<sup><xref ref-type="table-fn" rid="table2fn9">i</xref></sup></td><td align="left" valign="top">0.655<sup><xref ref-type="table-fn" rid="table2fn10">j</xref></sup></td><td align="left" valign="top">0.473<sup><xref ref-type="table-fn" rid="table2fn9">i</xref></sup></td><td align="left" valign="top">1</td><td align="left" valign="top"/></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content><italic>P</italic> value</td><td align="left" valign="top">.10</td><td align="left" valign="top">.25</td><td align="left" valign="top">.15</td><td align="left" valign="top">.32</td><td align="left" valign="top">.047</td><td align="left" valign="top">.02</td><td align="left" valign="top">.003</td><td align="left" valign="top">.047</td><td align="left" valign="top">&#x2014;</td><td align="left" valign="top"/></tr><tr><td align="left" valign="top" colspan="11">Wayfinding time</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content><italic>r</italic></td><td align="left" valign="top">&#x2212;0.549<sup><xref ref-type="table-fn" rid="table2fn9">i</xref></sup></td><td align="left" valign="top">&#x2212;0.458<sup><xref ref-type="table-fn" rid="table2fn9">i</xref></sup></td><td align="left" valign="top">&#x2212;0.588<sup><xref ref-type="table-fn" rid="table2fn9">i</xref></sup></td><td align="left" valign="top">&#x2212;0.296</td><td align="left" valign="top">0.716<sup><xref ref-type="table-fn" rid="table2fn10">j</xref></sup></td><td align="left" valign="top">&#x2212;0.751<sup><xref ref-type="table-fn" rid="table2fn7">g</xref></sup></td><td align="left" valign="top">0.426</td><td align="left" valign="top">0.216</td><td align="left" valign="top">0.920<sup><xref ref-type="table-fn" rid="table2fn7">g</xref></sup></td><td align="left" valign="top">1</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content><italic>P</italic> value</td><td align="left" valign="top">.02</td><td align="left" valign="top">.06</td><td align="left" valign="top">.01</td><td align="left" valign="top">.23</td><td align="left" valign="top">.002</td><td align="left" valign="top">&#x003C;.001</td><td align="left" valign="top">.08</td><td align="left" valign="top">.39</td><td align="left" valign="top">&#x003C;.001</td><td align="left" valign="top">&#x2014;</td></tr><tr><td align="left" valign="top" colspan="11">Navigation efficiency</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content><italic>r</italic></td><td align="left" valign="top">0.495<sup><xref ref-type="table-fn" rid="table2fn9">i</xref></sup></td><td align="left" valign="top">0.445</td><td align="left" valign="top">0.658<sup><xref ref-type="table-fn" rid="table2fn10">j</xref></sup></td><td align="left" valign="top">0.206</td><td align="left" valign="top">&#x2212;0.761<sup><xref ref-type="table-fn" rid="table2fn10">j</xref></sup></td><td align="left" valign="top">0.848<sup><xref ref-type="table-fn" rid="table2fn7">g</xref></sup></td><td align="left" valign="top">&#x2212;0.067</td><td align="left" valign="top">0.129</td><td align="left" valign="top">&#x2212;0.709<sup><xref ref-type="table-fn" rid="table2fn10">j</xref></sup></td><td align="left" valign="top">&#x2212;0.913<sup><xref ref-type="table-fn" rid="table2fn7">g</xref></sup></td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content><italic>P</italic> value</td><td align="left" valign="top">.04</td><td align="left" valign="top">.06</td><td align="left" valign="top">.003</td><td align="left" valign="top">.41</td><td align="left" valign="top">.001</td><td align="left" valign="top">&#x003C;.001</td><td align="left" valign="top">.79</td><td align="left" valign="top">.61</td><td align="left" valign="top">.001</td><td align="left" valign="top">&#x003C;.001</td></tr></tbody></table><table-wrap-foot><fn id="table2fn1"><p><sup>a</sup>MoCA: Montreal Cognitive Assessment.</p></fn><fn id="table2fn2"><p><sup>b</sup>MMSE: Mini-Mental State Examination.</p></fn><fn id="table2fn3"><p><sup>c</sup>ROCF-C: Copy condition of the Rey-Osterrieth Complex Figure.</p></fn><fn id="table2fn4"><p><sup>d</sup>ROCF-D: delayed recall condition of the Rey-Osterrieth Complex Figure.</p></fn><fn id="table2fn5"><p><sup>e</sup>TMTA: Part A of the Trail Making Test.</p></fn><fn id="table2fn6"><p><sup>f</sup>MRT: Mental Rotation Test.</p></fn><fn id="table2fn7"><p><sup>g</sup><italic>P</italic>&#x003C;.001</p></fn><fn id="table2fn8"><p><sup>h</sup>Not applicable.</p></fn><fn id="table2fn9"><p><sup>i</sup><italic>P</italic>&#x003C;.05</p></fn><fn id="table2fn10"><p><sup>j</sup><italic>P</italic>&#x003C;.01.</p></fn></table-wrap-foot></table-wrap></sec><sec id="s3-3"><title>Analysis of Age-Related Decline in Wayfinding Ability</title><p>The Mann-Whitney <italic>U</italic> test was used to assess differences in navigation performance between age groups (<xref ref-type="fig" rid="figure6">Figure 6</xref>). Compared to young participants, older adults, who exhibited significantly lower cognitive ability, demonstrated significantly longer navigation time (<italic>z</italic>=&#x2212;4.771, <italic>P</italic>&#x003C;.001) and wayfinding times (<italic>z</italic>=&#x2212;4.315, <italic>P</italic>&#x003C;.001), as well as reduced navigation efficiency (<italic>z</italic>=&#x2212;4.285, <italic>P</italic>&#x003C;.001).</p><fig position="float" id="figure6"><label>Figure 6.</label><caption><p>Box plots illustrating differences in navigation performance between older and younger participants.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="games_v14i1e83128_fig06.png"/></fig><p>The above metrics reflect participants&#x2019; overall wayfinding performance. However, navigation behavior is not solely influenced by internal factors such as cognitive ability. It also varies depending on the location of participants within the VE. <xref ref-type="fig" rid="figure7">Figure 7A</xref> illustrates differences in the navigation process across age groups during the formal experiment. To investigate this, participants&#x2019; spatial positions are mapped onto the shortest navigation routes (excluding trials involving route errors), and the time spent per unit distance is calculated for each individual. Based on these data, we plot curves of navigation time and navigation efficiency for each age group. Furthermore, a permutation test is conducted on navigation efficiency within each unit distance between older and young adults. The shaded regions on <xref ref-type="fig" rid="figure7">Figure 7B</xref> correspond to areas where significant differences are observed (<italic>P</italic>&#x003C;.05) in <xref ref-type="fig" rid="figure7">Figure 7A</xref>, revealing spatial zones where age-related differences in navigation behavior are most pronounced. These results highlight age-related differences in spatial behavior and reinforce the utility of these behavioral metrics as sensitive indicators of cognitive function.</p><fig position="float" id="figure7"><label>Figure 7.</label><caption><p>Comparison of the navigation process between older and younger participants during the formal experiment. (A) The x-axis represents the normalized value along the shortest distance across 5 consecutive trials, and the y-axis shows both navigation time and navigation efficiency. The colored bar at the bottom corresponds to the target color for each trial, and the labels R1, R2, ... indicate the axial lines that constitute the shortest navigation route for that trial. (B) Pink-shaded areas indicate route segments where significant differences in navigation efficiency were found between older and younger participants, as determined by permutation testing (<italic>P</italic>&#x003C;.05).</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="games_v14i1e83128_fig07.png"/></fig></sec><sec id="s3-4"><title>Effects of Environmental Structure on Navigation Behavior in Different Age Groups</title><p>The influence of environmental characteristics on navigation performance is analyzed from the perspective of spatial configuration using both AMA and VGA. Between-group comparisons revealed significant age-related differences in several experienced spatial measures. Older adults showed higher EAI (<italic>z</italic>=&#x2013;2.43, <italic>P</italic>=.01) and EVI (<italic>t</italic>=2.48, <italic>P</italic>=.02), whereas young adults exhibited higher EMAD (<italic>z</italic>=&#x2013;2.13, <italic>P</italic>=.03) and EMVD (<italic>t</italic>=&#x2013;2.49, <italic>P</italic>=.02). No significant group differences were found for experienced axial connectivity and experienced visual connectivity.</p><p>Correlation analyses further revealed distinct age-specific associations between navigation efficiency and experienced metrics. Among older adults, navigation efficiency showed a significant negative correlation with EVI (<italic>r</italic>=&#x2013;0.48, <italic>P</italic>=.04) and a positive correlation with EMVD (<italic>r</italic>=0.49, <italic>P</italic>=.03). In contrast, young adults demonstrated strong associations with axial properties, with navigation efficiency negatively correlated with EAI (<italic>r</italic>=&#x2013;0.64, <italic>P</italic>=.005) and positively correlated with EMAD (<italic>r</italic>=.72, <italic>P</italic>&#x003C;0.001). Young adults also exhibited a significant positive relationship between navigation efficiency and EMVD (<italic>r</italic>=0.57, <italic>P</italic>=.01).</p></sec></sec><sec id="s4" sec-type="discussion"><title>Discussion</title><sec id="s4-1"><title>Summary and Explanation of Findings</title><p>This study investigated how cognitive abilities and environmental structure jointly shape navigational performance in a virtual navigation game across different age groups. Consistent with prior research, our findings demonstrate that performance in navigation games, particularly navigation efficiency, is strongly influenced by individual cognitive capacity, as well as the spatial configuration of the environment. These results underscore the relevance of navigation behavior as an indicator of cognitive status and highlight the potential of VR-based navigation tasks as effective tools for cognitive assessment. Moreover, the insights gained from this work also provide valuable guidance for future urban design and navigation-game development.</p><p>The behavioral metrics used in this study, including traversal distance, number of errors, and navigation time, are widely used in human navigation research [<xref ref-type="bibr" rid="ref34">34</xref>,<xref ref-type="bibr" rid="ref60">60</xref>]. Spatiotemporal representation has long been essential for analyzing navigation behavior, and the present study extends conventional approaches by classifying participants&#x2019; navigation states using an individualized method. Navigation tasks are generally considered to consist of 2 main components [<xref ref-type="bibr" rid="ref61">61</xref>]: locomotion and wayfinding. Locomotion refers to the physical execution of movement in the space, whereas wayfinding is involved in determining and planning a route to a desired destination [<xref ref-type="bibr" rid="ref62">62</xref>]. Traditional approaches typically distinguish these components using fixed absolute speed thresholds [<xref ref-type="bibr" rid="ref47">47</xref>,<xref ref-type="bibr" rid="ref63">63</xref>]. To account for individual differences in navigation ability, we applied k-means clustering to each participant&#x2019;s instantaneous speed distribution to derive personalized relative speed thresholds. k-Means is a classical clustering algorithm that iteratively assigns data points to the nearest cluster center [<xref ref-type="bibr" rid="ref64">64</xref>]. To assess the reliability of this classification, clustering solutions with different numbers of clusters (<italic>k</italic>) were evaluated across all younger and older participants using the sum of squared errors (SSE) [<xref ref-type="bibr" rid="ref65">65</xref>] and the silhouette coefficient [<xref ref-type="bibr" rid="ref66">66</xref>] (Figure S3 in <xref ref-type="supplementary-material" rid="app2">Multimedia Appendix 2</xref>). The SSE quantifies within-cluster variance by measuring the squared distance between each data point and its assigned cluster center, whereas the silhouette coefficient provides a complementary assessment of clustering quality by comparing consistency within clusters with separation between other clusters [<xref ref-type="bibr" rid="ref67">67</xref>]. When <italic>k</italic>=3, the SSE curve showed a clear inflection point, indicating a markedly diminished rate of decrease in within-cluster variance. Meanwhile, the silhouette coefficient remained relatively high and exhibited more stable variance, suggesting an optimal balance between cluster cohesion and separation. In addition to these statistical indicators, the three clusters aligned well with meaningful navigation states observed in participants&#x2019; movement patterns (Figure S2 in <xref ref-type="supplementary-material" rid="app3">Multimedia Appendix 3</xref>): (1) low-speed hesitation or searching, (2) medium-speed transitional adjustments, and (3) high-speed goal-directed movement. Taken together, both the quantitative metrics and the behavioral interpretability support the individualized clustering approach, which enabled the effective categorization of each sampled time point into low-, medium-, or high-speed states, thereby identifying regions within the environment where participants associated with wayfinding, transition, or continuous movement.</p><p>Effective spatial navigation relies on the integration of multiple abilities, including visual perception, spatial orientation, learning, and memory [<xref ref-type="bibr" rid="ref68">68</xref>]. Wayfinding encompasses all of the ways in which people orient themselves in physical space and navigate from place to place. As anticipated, wayfinding ability, as a cognitive element of the navigation process [<xref ref-type="bibr" rid="ref69">69</xref>], is effectively captured by parameters such as wayfinding time and navigation efficiency, both of which are shown to be sensitive indicators of cognitive capacity in the current study. As summarized in <xref ref-type="table" rid="table2">Table 2</xref>, cognitive scales including the MMSE, MoCA, ROCF-C, TMTA, and MRT exhibited robust correlations with these navigation metrics. Among these, the MoCA and MMSE are widely used for comprehensive cognitive assessment; the TMTA primarily reflects executive functioning; ROCF-C performance is sensitive to visuoconstructive and geometric processing abilities; and the MRT assesses mental rotation capacity. These cognitive functions collectively support navigation by contributing to processes such as initial route planning, continuous movement control, reorientation, and decision-making. SG-based VR navigation tasks elicit continuous, ecologically relevant behavior under dynamic task constraints, enabling the simultaneous assessment of multiple cognitive functions within an integrated behavioral context. Quantifying wayfinding performance provides an indirect yet meaningful reflection of an individual&#x2019;s cognitive functioning [<xref ref-type="bibr" rid="ref34">34</xref>]. Longer wayfinding times and lower navigation efficiency were associated with poorer cognitive performance in our older participants. Taken together, these findings highlight the potential of well-designed navigation games to serve as effective tools for cognitive assessment [<xref ref-type="bibr" rid="ref70">70</xref>].</p><p>Human navigation in outdoor environments involves a series of distinctive behaviors, such as retracing steps, hesitation, and reorientation [<xref ref-type="bibr" rid="ref71">71</xref>]. These behaviors become particularly evident in complex spatial settings where individuals must process environmental cues and make directional decisions. Wayfinding time serves as a key indicator for revealing navigation uncertainty, effectively pinpointing locations where external environmental factors impact individuals&#x2019; sense of directional uncertainty (<xref ref-type="fig" rid="figure4">Figure 4</xref>). We observe that prolonged wayfinding times were predominantly concentrated in the turning areas of the path and the starting or ending locations of each trial. These highlighted areas in the figure represent regions where participants are more likely to encounter decision-making bottlenecks, experiencing uncertainty in selecting the correct path [<xref ref-type="bibr" rid="ref72">72</xref>]. Furthermore, integrating the analysis in <xref ref-type="fig" rid="figure6">Figure 6</xref>, we find significant differences in navigation performance between older and younger participants in certain regions.</p><p>Synthesizing findings from previous research, we propose that these results may be explained by age-related differences in navigation strategy preferences. Human navigation relies on 2 primary strategies: allocentric (survey-based) and egocentric (route-based). Younger adults tend to rely more on the allocentric strategy [<xref ref-type="bibr" rid="ref18">18</xref>] and show greater flexibility in switching between allocentric and egocentric frames of reference [<xref ref-type="bibr" rid="ref73">73</xref>]. However, aging is associated with a decline in allocentric navigation abilities, accompanied by a preferential shift toward egocentric strategies [<xref ref-type="bibr" rid="ref18">18</xref>,<xref ref-type="bibr" rid="ref21">21</xref>,<xref ref-type="bibr" rid="ref22">22</xref>], which may contribute to less efficient route planning and increased reliance on familiar or visually accessible paths. These strategy differences align closely with the behavioral patterns observed in our study. Young participants adopted a more proactive and goal-directed approach: they typically scanned the environment at the beginning of each trial, extracted relevant spatial cues, and formed a planned route before moving. In contrast, older participants showed a stimulus-response navigation pattern [<xref ref-type="bibr" rid="ref74">74</xref>,<xref ref-type="bibr" rid="ref75">75</xref>], making decisions only upon arriving at intersections rather than planning ahead. This led to hesitation at nearly every decision point (<xref ref-type="fig" rid="figure6">Figure 6B</xref>), reflecting greater reliance on immediate cues and reduced anticipation of upcoming turns. Consequently, participants spent more time searching for landmarks or other environmental cues, leading to decreased navigation performance [<xref ref-type="bibr" rid="ref76">76</xref>,<xref ref-type="bibr" rid="ref77">77</xref>]. Such behaviors are consistent with age-related declines in spatial working memory and executive function [<xref ref-type="bibr" rid="ref78">78</xref>], which limit their ability to efficiently integrate spatial cues.</p><p>To explore how environmental structure influences navigation across age groups, we used line-based (AMA) and grid-based (VGA) analytic approaches. Both methods quantify spatial configuration and offer complementary perspectives on how spatial structure shapes navigation behavior. Within this framework, integration serves as a core space syntax metric capturing the degree to which a path is connected to its surrounding environment [<xref ref-type="bibr" rid="ref13">13</xref>,<xref ref-type="bibr" rid="ref29">29</xref>]. In this study, experienced integration reflects the actual spatial integration values of the segments that participants traversed, thereby characterizing the dynamic interplay between environmental structure and navigational behavior. Older adults exhibited significantly higher EAI and EVI than young adults, suggesting a tendency to navigate toward more integrated areas rather than taking shorter or more direct routes. Prior studies have shown that spending more time in highly integrated areas facilitates the formation of accurate cognitive maps [<xref ref-type="bibr" rid="ref13">13</xref>]. With aging, older adults require more time to form cognitive maps and encode spatial information [<xref ref-type="bibr" rid="ref79">79</xref>], which may partly explain their tendency to remain within highly integrated axial regions during navigation. From another perspective, although integration is the normalization form of mean depth [<xref ref-type="bibr" rid="ref25">25</xref>], the two measures capture different aspects of spatial structure. Mean depth quantifies the average topological steps from a location to all others, with higher values indicating deeper and harder-to-reach positions. Older adults showed lower EMAD and EMVD values, suggesting a tendency to remain within relatively shallower, more accessible areas.</p><p>In our navigation game, all participants were instructed to reach the target location as quickly as possible. By comparing each group&#x2019;s experienced metrics with those of the shortest path (<xref ref-type="fig" rid="figure8">Figure 8A</xref>), we found that the experienced metrics of younger participants closely matched the optimal path. This pattern is consistent with behavioral observations showing that young participants, driven by the goal of rapid completion, tended to select more direct and efficient routes. Correlation analyses further support these environment-mediated navigation patterns. Among young adults, EAI was negatively correlated with navigation efficiency, indicating that more movement through globally integrated axial lines was associated with increased wayfinding behaviors (eg, hesitation, pausing, reorientation), thereby reducing overall efficiency. For older adults, EVI showed a significant negative correlation with navigation efficiency, suggesting that more extensive traversal within visually integrated regions corresponded to heightened hesitation or searching behaviors. The results for EMAD and EMVD exhibited similar trends: achieving higher navigation efficiency and shorter routes requires participants to strategically decide whether to traverse spatially advantageous or disadvantageous areas. Younger adults tended to accept navigating through deeper or less accessible regions to optimize route efficiency [<xref ref-type="bibr" rid="ref56">56</xref>], whereas older adults were more likely to remain within visually accessible areas, even when these were not part of the shortest path.</p><fig position="float" id="figure8"><label>Figure 8.</label><caption><p>Experienced metrics across age groups and their associations with navigation efficiency. (A) Group differences in 6 experienced spatial metrics, experienced axial integration (EAI), experienced axial connectivity, experienced mean axial depth (EMAD), experienced visual integration (EVI), experienced visual connectivity, and experienced visual mean depth (EVMD) are shown for older adults (dark blue) and young adults (light blue). The brown dashed line represents the experienced metric obtained from the shortest path. (B) Correlations between navigation efficiency and 4 key experienced metrics (EAI, EMAD, EVI, and EVMD) are presented separately for older adults (dark blue) and young adults (light blue). Pearson correlations are denoted by (p) and Spearman correlations by (s).</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="games_v14i1e83128_fig08.png"/></fig><p>This subjective tendency in navigational behavior indicates that participants do not passively respond only to environmental structure; rather, their navigational choices also reflect intentional, strategy-driven preferences [<xref ref-type="bibr" rid="ref13">13</xref>]. Such findings point to a bidirectional relationship between spatial configuration and navigation behavior, where environmental structure shapes movement patterns, and individuals&#x2019; strategic preferences, in turn, influence the parts of the environment they traverse. Understanding whether environmental properties or individual behavior play the dominant role in guiding navigation will require further empirical investigation.</p><p>Additionally, our findings extend this literature by demonstrating age-specific patterns in the relationship between experienced metrics and navigation efficiency. Navigation efficiency in younger adults was strongly associated with line-based (AMA) metrics such as EAI and EMAD, whereas older adults showed stronger associations with grid-based (VGA) metrics, such as EVI and EMVD. These differences can be meaningfully interpreted through the lens of allocentric and egocentric navigation strategies. Younger adults tend to rely more on allocentric representations, enabling them to rapidly extract structural information from the environment, integrate axial connectivity, and flexibly switch between global structure and local visual information. This cognitive flexibility allows them to traverse deeper or less accessible areas. In contrast, aging is associated with an increased reliance on egocentric navigation. Older adults become more dependent on the visual field and immediate cues, which aligns with their stronger correlation between navigation efficiency and visual-based experienced metrics (EVI and EMVD). Their tendency to remain within visually accessible, highly integrated regions suggests that they prioritize environments that offer clear visual guidance and reduced cognitive demands [<xref ref-type="bibr" rid="ref14">14</xref>].</p><p>Together, space syntax highlights the fundamental role of spatial configuration in shaping navigation behavior, and our findings further show that its influence varies with age, ultimately affecting navigation performance. These insights have important implications not only for developing more effective cognitive assessment tools but also for designing age-friendly urban environments that support safe, efficient, and accessible navigation. Strengthening the coordination and integration of street space is essential for urban street-design guidelines and road-network optimization [<xref ref-type="bibr" rid="ref80">80</xref>]. Through axial and grid-based analyses, space syntax provides a powerful framework for identifying high integration. Such information can guide the placement of landmarks, directional signage, and other environmental cues that facilitate cognitive map formation and enhance navigability for diverse user groups. Moreover, because spatial integration is closely tied to navigation difficulty [<xref ref-type="bibr" rid="ref29">29</xref>], purposeful modulation of integration levels can help calibrate the cognitive demands of navigation tasks. From an SG perspective, the majority of participants verbally reported a high level of immersion and task engagement, suggesting that the SG-based framework effectively supported sustained interaction with the spatial environment. By integrating principles of spatial configuration into an immersive VR navigation game, this approach may provide promising opportunities for the development of gamified tools for cognitive assessment, training, and rehabilitation, particularly for individuals with spatial or cognitive impairments.</p></sec><sec id="s4-2"><title>Limitations and Future Work</title><p>This research highlights the critical role of cognitive abilities and spatial configuration in shaping navigation behavior. However, several limitations should be acknowledged. First, the participant sample may not fully represent the broader population. The young participants were primarily between 20 and 30 years of age, whereas the older group ranged from 65 to 85 years, leaving a substantial gap between ages 30 and 65 years, which was not included. This omission limits the generalizability of the findings, as individuals in midlife may exhibit distinct navigation patterns and cognitive characteristics that were not captured in the current analysis [<xref ref-type="bibr" rid="ref72">72</xref>]. Second, the individualized k-means clustering approach adopted in this study is particularly suited to navigation games characterized by relatively stable movement speeds. It may be less appropriate for tasks involving pronounced, irregular, or continuous speed transitions, in which clear and interpretable cluster boundaries are difficult to establish. Moreover, although a 3-cluster solution showed good behavioral interpretability within the current navigation game, the generalizability of this clustering scheme to other navigation games or more complex real-world environments remains to be validated in future studies. Third, discrepancies between virtual and real-world navigation are unavoidable [<xref ref-type="bibr" rid="ref76">76</xref>]. The absence of environmental features such as directional signs, combined with the lack of natural physical movement, may influence how spatial information is encoded, updated, and retrieved [<xref ref-type="bibr" rid="ref81">81</xref>]. Finally, the study did not explicitly account for the influence of demographic factors such as gender, educational background, or geographic and cultural experience. Ignoring these variables could obscure important individual differences and interactions relevant to wayfinding performance.</p><p>Future research should aim to address these limitations by recruiting a more demographically diverse participant sample, including middle-aged individuals, and by designing VEs with greater ecological validity. Incorporating navigation methods that more closely approximate natural physical movement may yield richer behavioral data and uncover additional nuances in navigational performance. Additionally, systematically examining the role of experienced metrics could provide deeper insights into urban design, spatial planning, and navigation-aid development in both virtual and real-world contexts. From a rehabilitation perspective, the game demonstrates promising potential for virtual navigation tasks as a controlled and adaptable platform for personalized cognitive assessment and training. Further research could investigate the incorporation of adaptive task difficulty and longitudinal monitoring into VR-based rehabilitation, thereby supporting the maintenance or recovery of spatial and cognitive functions.</p></sec><sec id="s4-3"><title>Conclusions</title><p>This study provides quantitative evidence for how environmental configuration and cognitive abilities jointly shape navigational behavior in VEs. By applying an individualized k-means clustering approach, we extracted key behavioral indicators, such as wayfinding time and navigation efficiency, that proved sensitive to cognitive decline. This method offers a more personalized and potentially more effective framework for assessing cognitive function through navigation tasks. Through spatial analysis, we further identified age-specific regions of wayfinding difficulty and quantified the extent to which environmental structure influences navigation behavior. These findings reveal distinct navigation patterns between younger and older adults and underscore the value of integrating behavioral analytics with spatial-syntax metrics. Overall, the results lay an important foundation for future applications in clinical cognitive assessment and rehabilitation, as well as in the development of age-friendly urban environments.</p></sec></sec></body><back><notes><sec><title>Funding</title><p>This work was supported by the National Nature Science Foundation of China (12372326, 12472302, T2288101) and the Beijing Municipal Natural Science Foundation (L244020).</p></sec></notes><fn-group><fn fn-type="con"><p>XL and ST contributed equally as co-corresponding authors. ST can be contacted by email at shantian@buaa.edu.cn or by phone at +86-15010725395.</p></fn><fn fn-type="conflict"><p>None declared.</p></fn></fn-group><glossary><title>Abbreviations</title><def-list><def-item><term id="abb1">AMA</term><def><p>axial map analysis</p></def></def-item><def-item><term id="abb2">EAI</term><def><p>experienced axial integration</p></def></def-item><def-item><term id="abb3">EMAD</term><def><p>experienced mean axial depth</p></def></def-item><def-item><term id="abb4">EVI</term><def><p>experienced visual integration</p></def></def-item><def-item><term id="abb5">MMSE</term><def><p>Mini-Mental State Examination</p></def></def-item><def-item><term id="abb6">MoCA</term><def><p>Montreal Cognitive Assessment</p></def></def-item><def-item><term id="abb7">MRT</term><def><p>Mental Rotation Test</p></def></def-item><def-item><term id="abb8">ROCF-C</term><def><p>Copy condition of the Rey-Osterrieth Complex Figure</p></def></def-item><def-item><term id="abb9">SG</term><def><p>serious game</p></def></def-item><def-item><term id="abb10">SSE</term><def><p>sum of squared errors</p></def></def-item><def-item><term id="abb11">TMTA</term><def><p>Part A of the Trail Making Test</p></def></def-item><def-item><term id="abb12">VE</term><def><p>virtual environment</p></def></def-item><def-item><term id="abb13">VGA</term><def><p>visibility graph analysis</p></def></def-item><def-item><term id="abb14">VR</term><def><p>virtual reality</p></def></def-item></def-list></glossary><ref-list><title>References</title><ref id="ref1"><label>1</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Liu</surname><given-names>B</given-names> </name><name name-style="western"><surname>Dong</surname><given-names>WH</given-names> </name><name name-style="western"><surname>Zhan</surname><given-names>ZC</given-names> </name><name name-style="western"><surname>Wang</surname><given-names>SK</given-names> </name><name name-style="western"><surname>Meng</surname><given-names>LQ</given-names> </name></person-group><article-title>Differences in the gaze behaviours of pedestrians navigating between regular and irregular road patterns</article-title><source>ISPRS Int J Geo-Inf</source><year>2020</year><month>01</month><volume>9</volume><issue>1</issue><fpage>45</fpage><pub-id pub-id-type="doi">10.3390/ijgi9010045</pub-id></nlm-citation></ref><ref id="ref2"><label>2</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Lester</surname><given-names>AW</given-names> </name><name name-style="western"><surname>Moffat</surname><given-names>SD</given-names> </name><name name-style="western"><surname>Wiener</surname><given-names>JM</given-names> </name><name name-style="western"><surname>Barnes</surname><given-names>CA</given-names> </name><name name-style="western"><surname>Wolbers</surname><given-names>T</given-names> </name></person-group><article-title>The aging navigational system</article-title><source>Neuron</source><year>2017</year><month>08</month><day>30</day><volume>95</volume><issue>5</issue><fpage>1019</fpage><lpage>1035</lpage><pub-id pub-id-type="doi">10.1016/j.neuron.2017.06.037</pub-id><pub-id pub-id-type="medline">28858613</pub-id></nlm-citation></ref><ref id="ref3"><label>3</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Moffat</surname><given-names>SD</given-names> </name></person-group><article-title>Aging and spatial navigation: what do we know and where do we go?</article-title><source>Neuropsychol Rev</source><year>2009</year><month>12</month><volume>19</volume><issue>4</issue><fpage>478</fpage><lpage>489</lpage><pub-id pub-id-type="doi">10.1007/s11065-009-9120-3</pub-id><pub-id pub-id-type="medline">19936933</pub-id></nlm-citation></ref><ref id="ref4"><label>4</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Guariglia</surname><given-names>CC</given-names> </name><name name-style="western"><surname>Nitrini</surname><given-names>R</given-names> </name></person-group><article-title>Topographical disorientation in Alzheimer&#x2019;s disease</article-title><source>Arq Neuropsiquiatr</source><year>2009</year><month>12</month><volume>67</volume><issue>4</issue><fpage>967</fpage><lpage>972</lpage><pub-id pub-id-type="doi">10.1590/s0004-282x2009000600001</pub-id><pub-id pub-id-type="medline">20069202</pub-id></nlm-citation></ref><ref id="ref5"><label>5</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Iaria</surname><given-names>G</given-names> </name></person-group><article-title>Developmental topographical disorientation: lost every day</article-title><source>Lancet Neurol</source><year>2013</year><month>08</month><volume>12</volume><issue>8</issue><fpage>745</fpage><pub-id pub-id-type="doi">10.1016/S1474-4422(13)70133-9</pub-id></nlm-citation></ref><ref id="ref6"><label>6</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Igarashi</surname><given-names>KM</given-names> </name></person-group><article-title>Entorhinal cortex dysfunction in Alzheimer&#x2019;s disease</article-title><source>Trends Neurosci</source><year>2023</year><month>02</month><volume>46</volume><issue>2</issue><fpage>124</fpage><lpage>136</lpage><pub-id pub-id-type="doi">10.1016/j.tins.2022.11.006</pub-id><pub-id pub-id-type="medline">36513524</pub-id></nlm-citation></ref><ref id="ref7"><label>7</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Stangl</surname><given-names>M</given-names> </name><name name-style="western"><surname>Achtzehn</surname><given-names>J</given-names> </name><name name-style="western"><surname>Huber</surname><given-names>K</given-names> </name><name name-style="western"><surname>Dietrich</surname><given-names>C</given-names> </name><name name-style="western"><surname>Tempelmann</surname><given-names>C</given-names> </name><name name-style="western"><surname>Wolbers</surname><given-names>T</given-names> </name></person-group><article-title>Compromised grid-cell-like representations in old age as a key mechanism to explain age-related navigational deficits</article-title><source>Curr Biol</source><year>2018</year><month>04</month><day>2</day><volume>28</volume><issue>7</issue><fpage>1108</fpage><lpage>1115</lpage><pub-id pub-id-type="doi">10.1016/j.cub.2018.02.038</pub-id><pub-id pub-id-type="medline">29551413</pub-id></nlm-citation></ref><ref id="ref8"><label>8</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Kinkhabwala</surname><given-names>AA</given-names> </name><name name-style="western"><surname>Gu</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Aronov</surname><given-names>D</given-names> </name><name name-style="western"><surname>Tank</surname><given-names>DW</given-names> </name></person-group><article-title>Visual cue-related activity of cells in the medial entorhinal cortex during navigation in virtual reality</article-title><source>Elife</source><year>2020</year><month>03</month><day>9</day><volume>9</volume><fpage>e43140</fpage><pub-id pub-id-type="doi">10.7554/eLife.43140</pub-id><pub-id pub-id-type="medline">32149601</pub-id></nlm-citation></ref><ref id="ref9"><label>9</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Zhang</surname><given-names>TT</given-names> </name><name name-style="western"><surname>Lian</surname><given-names>ZF</given-names> </name><name name-style="western"><surname>Xu</surname><given-names>YN</given-names> </name></person-group><article-title>Combining GPS and space syntax analysis to improve understanding of visitor temporal&#x2013;spatial behaviour: a case study of the Lion Grove in China</article-title><source>Landsc Res</source><year>2020</year><month>05</month><day>18</day><volume>45</volume><issue>4</issue><fpage>534</fpage><lpage>546</lpage><pub-id pub-id-type="doi">10.1080/01426397.2020.1730775</pub-id></nlm-citation></ref><ref id="ref10"><label>10</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Hegarty</surname><given-names>M</given-names> </name><name name-style="western"><surname>He</surname><given-names>CXY</given-names> </name><name name-style="western"><surname>Boone</surname><given-names>AP</given-names> </name><name name-style="western"><surname>Yu</surname><given-names>SY</given-names> </name><name name-style="western"><surname>Jacobs</surname><given-names>EG</given-names> </name><name name-style="western"><surname>Chrastil</surname><given-names>ER</given-names> </name></person-group><article-title>Understanding differences in wayfinding strategies</article-title><source>Top Cogn Sci</source><year>2023</year><month>01</month><volume>15</volume><issue>1</issue><fpage>102</fpage><lpage>119</lpage><pub-id pub-id-type="doi">10.1111/tops.12592</pub-id><pub-id pub-id-type="medline">34973064</pub-id></nlm-citation></ref><ref id="ref11"><label>11</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Kafri</surname><given-names>M</given-names> </name><name name-style="western"><surname>Weiss</surname><given-names>PL</given-names> </name><name name-style="western"><surname>Zeilig</surname><given-names>G</given-names> </name><name name-style="western"><surname>Bondi</surname><given-names>M</given-names> </name><name name-style="western"><surname>Baum-Cohen</surname><given-names>I</given-names> </name><name name-style="western"><surname>Kizony</surname><given-names>R</given-names> </name></person-group><article-title>Performance in complex life situations: effects of age, cognition, and walking speed in virtual versus real life environments</article-title><source>J Neuroeng Rehabil</source><year>2021</year><month>02</month><day>8</day><volume>18</volume><issue>1</issue><fpage>30</fpage><pub-id pub-id-type="doi">10.1186/s12984-021-00830-6</pub-id><pub-id pub-id-type="medline">33557894</pub-id></nlm-citation></ref><ref id="ref12"><label>12</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Coutrot</surname><given-names>A</given-names> </name><name name-style="western"><surname>Manley</surname><given-names>E</given-names> </name><name name-style="western"><surname>Goodroe</surname><given-names>S</given-names> </name><etal/></person-group><article-title>Entropy of city street networks linked to future spatial navigation ability</article-title><source>Nature New Biol</source><year>2022</year><month>04</month><day>7</day><volume>604</volume><issue>7904</issue><fpage>104</fpage><lpage>110</lpage><pub-id pub-id-type="doi">10.1038/s41586-022-04486-7</pub-id></nlm-citation></ref><ref id="ref13"><label>13</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Brunec</surname><given-names>IK</given-names> </name><name name-style="western"><surname>Nantais</surname><given-names>MM</given-names> </name><name name-style="western"><surname>Sutton</surname><given-names>JE</given-names> </name><name name-style="western"><surname>Epstein</surname><given-names>RA</given-names> </name><name name-style="western"><surname>Newcombe</surname><given-names>NS</given-names> </name></person-group><article-title>Exploration patterns shape cognitive map learning</article-title><source>Cognition</source><year>2023</year><month>04</month><volume>233</volume><fpage>105360</fpage><pub-id pub-id-type="doi">10.1016/j.cognition.2022.105360</pub-id><pub-id pub-id-type="medline">36549130</pub-id></nlm-citation></ref><ref id="ref14"><label>14</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Pagkratidou</surname><given-names>M</given-names> </name><name name-style="western"><surname>Galati</surname><given-names>A</given-names> </name><name name-style="western"><surname>Avraamides</surname><given-names>M</given-names> </name></person-group><article-title>Do environmental characteristics predict spatial memory about unfamiliar environments?</article-title><source>Spat Cogn Comput</source><year>2020</year><month>01</month><day>2</day><volume>20</volume><issue>1</issue><fpage>1</fpage><lpage>32</lpage><pub-id pub-id-type="doi">10.1080/13875868.2019.1676248</pub-id></nlm-citation></ref><ref id="ref15"><label>15</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Merhav</surname><given-names>M</given-names> </name><name name-style="western"><surname>Fisher-Gewirtzman</surname><given-names>D</given-names> </name></person-group><article-title>How pathways&#x2019; configuration impacts wayfinding in young and older adults</article-title><source>J Environ Psychol</source><year>2023</year><month>09</month><volume>90</volume><fpage>102065</fpage><pub-id pub-id-type="doi">10.1016/j.jenvp.2023.102065</pub-id></nlm-citation></ref><ref id="ref16"><label>16</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Krupic</surname><given-names>J</given-names> </name><name name-style="western"><surname>Bauza</surname><given-names>M</given-names> </name><name name-style="western"><surname>Burton</surname><given-names>S</given-names> </name><name name-style="western"><surname>Barry</surname><given-names>C</given-names> </name><name name-style="western"><surname>O&#x2019;Keefe</surname><given-names>J</given-names> </name></person-group><article-title>Grid cell symmetry is shaped by environmental geometry</article-title><source>Nature New Biol</source><year>2015</year><month>02</month><day>12</day><volume>518</volume><issue>7538</issue><fpage>232</fpage><lpage>235</lpage><pub-id pub-id-type="doi">10.1038/nature14153</pub-id><pub-id pub-id-type="medline">25673417</pub-id></nlm-citation></ref><ref id="ref17"><label>17</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Epstein</surname><given-names>RA</given-names> </name><name name-style="western"><surname>Patai</surname><given-names>EZ</given-names> </name><name name-style="western"><surname>Julian</surname><given-names>JB</given-names> </name><name name-style="western"><surname>Spiers</surname><given-names>HJ</given-names> </name></person-group><article-title>The cognitive map in humans: spatial navigation and beyond</article-title><source>Nat Neurosci</source><year>2017</year><month>10</month><day>26</day><volume>20</volume><issue>11</issue><fpage>1504</fpage><lpage>1513</lpage><pub-id pub-id-type="doi">10.1038/nn.4656</pub-id><pub-id pub-id-type="medline">29073650</pub-id></nlm-citation></ref><ref id="ref18"><label>18</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Rodgers</surname><given-names>MK</given-names> </name><name name-style="western"><surname>Sindone</surname><given-names>JA</given-names>  <suffix>III</suffix></name><name name-style="western"><surname>Moffat</surname><given-names>SD</given-names> </name></person-group><article-title>Effects of age on navigation strategy</article-title><source>Neurobiol Aging</source><year>2012</year><month>01</month><volume>33</volume><issue>1</issue><fpage>202</fpage><pub-id pub-id-type="doi">10.1016/j.neurobiolaging.2010.07.021</pub-id><pub-id pub-id-type="medline">20832911</pub-id></nlm-citation></ref><ref id="ref19"><label>19</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Tversky</surname><given-names>B</given-names> </name></person-group><article-title>Structures of mental spaces: how people think about space</article-title><source>Environ Behav</source><year>2003</year><volume>35</volume><issue>1</issue><fpage>66</fpage><lpage>80</lpage><pub-id pub-id-type="doi">10.1177/0013916502238865</pub-id></nlm-citation></ref><ref id="ref20"><label>20</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>B&#x00E9;cu</surname><given-names>M</given-names> </name><name name-style="western"><surname>Sheynikhovich</surname><given-names>D</given-names> </name><name name-style="western"><surname>Tatur</surname><given-names>G</given-names> </name><etal/></person-group><article-title>Age-related preference for geometric spatial cues during real-world navigation</article-title><source>Nat Hum Behav</source><year>2020</year><month>01</month><volume>4</volume><issue>1</issue><fpage>88</fpage><lpage>99</lpage><pub-id pub-id-type="doi">10.1038/s41562-019-0718-z</pub-id><pub-id pub-id-type="medline">31548677</pub-id></nlm-citation></ref><ref id="ref21"><label>21</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Harris</surname><given-names>MA</given-names> </name><name name-style="western"><surname>Wiener</surname><given-names>JM</given-names> </name><name name-style="western"><surname>Wolbers</surname><given-names>T</given-names> </name></person-group><article-title>Aging specifically impairs switching to an allocentric navigational strategy</article-title><source>Front Aging Neurosci</source><year>2012</year><volume>4</volume><issue>29</issue><fpage>29</fpage><pub-id pub-id-type="doi">10.3389/fnagi.2012.00029</pub-id><pub-id pub-id-type="medline">23125833</pub-id></nlm-citation></ref><ref id="ref22"><label>22</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>van der Ham</surname><given-names>IJM</given-names> </name><name name-style="western"><surname>Claessen</surname><given-names>MHG</given-names> </name></person-group><article-title>How age relates to spatial navigation performance: functional and methodological considerations</article-title><source>Ageing Res Rev</source><year>2020</year><month>03</month><volume>58</volume><fpage>101020</fpage><pub-id pub-id-type="doi">10.1016/j.arr.2020.101020</pub-id><pub-id pub-id-type="medline">31954190</pub-id></nlm-citation></ref><ref id="ref23"><label>23</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Hillier</surname><given-names>B</given-names> </name><name name-style="western"><surname>Hanson</surname><given-names>J</given-names> </name></person-group><source>The Social Logic of Space</source><year>1984</year><publisher-name>Cambridge University Press</publisher-name><pub-id pub-id-type="doi">10.1017/CBO9780511597237</pub-id><pub-id pub-id-type="other">9780521367844</pub-id></nlm-citation></ref><ref id="ref24"><label>24</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Jiang</surname><given-names>B</given-names> </name><name name-style="western"><surname>Claramunt</surname><given-names>C</given-names> </name><name name-style="western"><surname>Klarqvist</surname><given-names>B</given-names> </name></person-group><article-title>Integration of space syntax into GIS for modelling urban spaces</article-title><source>Int J Appl Earth Obs Geoinf</source><year>2000</year><volume>2</volume><issue>3-4</issue><fpage>161</fpage><lpage>171</lpage><pub-id pub-id-type="doi">10.1016/S0303-2434(00)85010-2</pub-id></nlm-citation></ref><ref id="ref25"><label>25</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Hillier</surname><given-names>B</given-names> </name></person-group><source>Space Is the Machine: A Configurational Theory of Architecture</source><year>1998</year><access-date>2026-03-09</access-date><publisher-name>Cambridge University Press</publisher-name><comment><ext-link ext-link-type="uri" xlink:href="https://discovery.ucl.ac.uk/id/eprint/3881/1/SITM.pdf">https://discovery.ucl.ac.uk/id/eprint/3881/1/SITM.pdf</ext-link></comment><pub-id pub-id-type="other">9780521645287</pub-id></nlm-citation></ref><ref id="ref26"><label>26</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Peponis</surname><given-names>J</given-names> </name><name name-style="western"><surname>Zimring</surname><given-names>C</given-names> </name><name name-style="western"><surname>Choi</surname><given-names>YK</given-names> </name></person-group><article-title>Finding the building in wayfinding</article-title><source>Environ Behav</source><year>1990</year><month>09</month><volume>22</volume><issue>5</issue><fpage>555</fpage><lpage>590</lpage><pub-id pub-id-type="doi">10.1177/0013916590225001</pub-id></nlm-citation></ref><ref id="ref27"><label>27</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>De Cock</surname><given-names>L</given-names> </name><name name-style="western"><surname>Van de Weghe</surname><given-names>N</given-names> </name><name name-style="western"><surname>Ooms</surname><given-names>K</given-names> </name><etal/></person-group><article-title>Linking the cognitive load induced by route instruction types and building configuration during indoor route guidance, a usability study in VR</article-title><source>Int J Geogr Inf Sci</source><year>2022</year><month>10</month><day>3</day><volume>36</volume><issue>10</issue><fpage>1978</fpage><lpage>2008</lpage><pub-id pub-id-type="doi">10.1080/13658816.2022.2032080</pub-id></nlm-citation></ref><ref id="ref28"><label>28</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Li</surname><given-names>R</given-names> </name><name name-style="western"><surname>Klippel</surname><given-names>A</given-names> </name></person-group><article-title>Wayfinding behaviors in complex buildings: the impact of environmental legibility and familiarity</article-title><source>Environ Behav</source><year>2016</year><month>04</month><volume>48</volume><issue>3</issue><fpage>482</fpage><lpage>510</lpage><pub-id pub-id-type="doi">10.1177/0013916514550243</pub-id></nlm-citation></ref><ref id="ref29"><label>29</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Yesiltepe</surname><given-names>D</given-names> </name><name name-style="western"><surname>Fern&#x00E1;ndez Velasco</surname><given-names>P</given-names> </name><name name-style="western"><surname>Coutrot</surname><given-names>A</given-names> </name><etal/></person-group><article-title>Entropy and a sub-group of geometric measures of paths predict the navigability of an environment</article-title><source>Cognition</source><year>2023</year><month>07</month><volume>236</volume><fpage>105443</fpage><pub-id pub-id-type="doi">10.1016/j.cognition.2023.105443</pub-id><pub-id pub-id-type="medline">37003236</pub-id></nlm-citation></ref><ref id="ref30"><label>30</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Neo</surname><given-names>JRJ</given-names> </name><name name-style="western"><surname>Won</surname><given-names>AS</given-names> </name><name name-style="western"><surname>Shepley</surname><given-names>MM</given-names> </name></person-group><article-title>Designing immersive virtual environments for human behavior research</article-title><source>Front Virtual Real</source><year>2021</year><month>03</month><volume>2</volume><fpage>603750</fpage><pub-id pub-id-type="doi">10.3389/frvir.2021.603750</pub-id></nlm-citation></ref><ref id="ref31"><label>31</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Slater</surname><given-names>M</given-names> </name><name name-style="western"><surname>Wilbur</surname><given-names>S</given-names> </name></person-group><article-title>A framework for immersive virtual environments (FIVE): speculations on the role of presence in virtual environments</article-title><source>Presence: Teleop Virtual Environ</source><year>1997</year><month>12</month><volume>6</volume><issue>6</issue><fpage>603</fpage><lpage>616</lpage><pub-id pub-id-type="doi">10.1162/pres.1997.6.6.603</pub-id></nlm-citation></ref><ref id="ref32"><label>32</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Checa</surname><given-names>D</given-names> </name><name name-style="western"><surname>Bustillo</surname><given-names>A</given-names> </name></person-group><article-title>A review of immersive virtual reality serious games to enhance learning and training</article-title><source>Multimed Tools Appl</source><year>2020</year><month>03</month><volume>79</volume><issue>9-10</issue><fpage>5501</fpage><lpage>5527</lpage><pub-id pub-id-type="doi">10.1007/s11042-019-08348-9</pub-id></nlm-citation></ref><ref id="ref33"><label>33</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Martinez</surname><given-names>ES</given-names> </name><name name-style="western"><surname>Wu</surname><given-names>AS</given-names> </name><name name-style="western"><surname>McMahan</surname><given-names>RP</given-names> </name></person-group><article-title>Research trends in virtual reality locomotion techniques</article-title><conf-name>2022 IEEE on Conference Virtual Reality and 3D User Interfaces (VR)</conf-name><conf-date>Mar 12-16, 2022</conf-date><pub-id pub-id-type="doi">10.1109/VR51125.2022.00046</pub-id></nlm-citation></ref><ref id="ref34"><label>34</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Martelli</surname><given-names>D</given-names> </name><name name-style="western"><surname>Prado</surname><given-names>A</given-names> </name><name name-style="western"><surname>Xia</surname><given-names>BX</given-names> </name><name name-style="western"><surname>Verghese</surname><given-names>J</given-names> </name><name name-style="western"><surname>Agrawal</surname><given-names>SK</given-names> </name></person-group><article-title>Development of a virtual floor maze test-effects of distal visual cues and correlations with executive function in healthy adults</article-title><source>IEEE Trans Neural Syst Rehabil Eng</source><year>2019</year><month>10</month><volume>27</volume><issue>10</issue><fpage>2229</fpage><lpage>2236</lpage><pub-id pub-id-type="doi">10.1109/TNSRE.2019.2938103</pub-id><pub-id pub-id-type="medline">31478863</pub-id></nlm-citation></ref><ref id="ref35"><label>35</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Roth</surname><given-names>D</given-names> </name><name name-style="western"><surname>Purps</surname><given-names>CF</given-names> </name><name name-style="western"><surname>Neumann</surname><given-names>WJ</given-names> </name></person-group><article-title>A virtual Morris water maze to study neurodegenarative disorders</article-title><conf-name>2020 IEEE International Symposium on Mixed and Augmented Reality Adjunct (ISMAR-Adjunct)</conf-name><conf-date>Nov 9-13, 2020</conf-date><pub-id pub-id-type="doi">10.1109/ISMAR-Adjunct51615.2020.00048</pub-id></nlm-citation></ref><ref id="ref36"><label>36</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Jacobs</surname><given-names>J</given-names> </name><name name-style="western"><surname>Weidemann</surname><given-names>CT</given-names> </name><name name-style="western"><surname>Miller</surname><given-names>JF</given-names> </name><etal/></person-group><article-title>Direct recordings of grid-like neuronal activity in human spatial navigation</article-title><source>Nat Neurosci</source><year>2013</year><month>09</month><volume>16</volume><issue>9</issue><fpage>1188</fpage><lpage>1190</lpage><pub-id pub-id-type="doi">10.1038/nn.3466</pub-id><pub-id pub-id-type="medline">23912946</pub-id></nlm-citation></ref><ref id="ref37"><label>37</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Irshad</surname><given-names>S</given-names> </name><name name-style="western"><surname>Perkis</surname><given-names>A</given-names> </name><name name-style="western"><surname>Azam</surname><given-names>W</given-names> </name></person-group><article-title>Wayfinding in virtual reality serious game: an exploratory study in the context of user perceived experiences</article-title><source>Appl Sci (Basel)</source><year>2021</year><volume>11</volume><issue>17</issue><fpage>7822</fpage><pub-id pub-id-type="doi">10.3390/app11177822</pub-id></nlm-citation></ref><ref id="ref38"><label>38</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>De Leeuw</surname><given-names>D</given-names> </name><name name-style="western"><surname>De Maeyer</surname><given-names>P</given-names> </name><name name-style="western"><surname>De Cock</surname><given-names>L</given-names> </name></person-group><article-title>A gamification-based approach on indoor wayfinding research</article-title><source>ISPRS Int J Geo-Inf</source><year>2020</year><volume>9</volume><issue>7</issue><fpage>423</fpage><pub-id pub-id-type="doi">10.3390/ijgi9070423</pub-id></nlm-citation></ref><ref id="ref39"><label>39</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Coughlan</surname><given-names>G</given-names> </name><name name-style="western"><surname>Coutrot</surname><given-names>A</given-names> </name><name name-style="western"><surname>Khondoker</surname><given-names>M</given-names> </name><name name-style="western"><surname>Minihane</surname><given-names>AM</given-names> </name><name name-style="western"><surname>Spiers</surname><given-names>H</given-names> </name><name name-style="western"><surname>Hornberger</surname><given-names>M</given-names> </name></person-group><article-title>Toward personalized cognitive diagnostics of at-genetic-risk Alzheimer&#x2019;s disease</article-title><source>Proc Natl Acad Sci U S A</source><year>2019</year><month>05</month><day>7</day><volume>116</volume><issue>19</issue><fpage>9285</fpage><lpage>9292</lpage><pub-id pub-id-type="doi">10.1073/pnas.1901600116</pub-id><pub-id pub-id-type="medline">31015296</pub-id></nlm-citation></ref><ref id="ref40"><label>40</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Coutrot</surname><given-names>A</given-names> </name><name name-style="western"><surname>Silva</surname><given-names>R</given-names> </name><name name-style="western"><surname>Manley</surname><given-names>E</given-names> </name><etal/></person-group><article-title>Global determinants of navigation ability</article-title><source>Curr Biol</source><year>2018</year><month>09</month><day>10</day><volume>28</volume><issue>17</issue><fpage>2861</fpage><lpage>2866.e4</lpage><pub-id pub-id-type="doi">10.1016/j.cub.2018.06.009</pub-id><pub-id pub-id-type="medline">30100340</pub-id></nlm-citation></ref><ref id="ref41"><label>41</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>West</surname><given-names>GL</given-names> </name><name name-style="western"><surname>Patai</surname><given-names>ZE</given-names> </name><name name-style="western"><surname>Coutrot</surname><given-names>A</given-names> </name><name name-style="western"><surname>Hornberger</surname><given-names>M</given-names> </name><name name-style="western"><surname>Bohbot</surname><given-names>VD</given-names> </name><name name-style="western"><surname>Spiers</surname><given-names>HJ</given-names> </name></person-group><article-title>Landmark-dependent navigation strategy declines across the human life-span: evidence from over 37,000 participants</article-title><source>J Cogn Neurosci</source><year>2023</year><month>03</month><day>1</day><volume>35</volume><issue>3</issue><fpage>452</fpage><lpage>467</lpage><pub-id pub-id-type="doi">10.1162/jocn_a_01956</pub-id><pub-id pub-id-type="medline">36603038</pub-id></nlm-citation></ref><ref id="ref42"><label>42</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Tragantzopoulou</surname><given-names>P</given-names> </name><name name-style="western"><surname>Giannouli</surname><given-names>V</given-names> </name></person-group><article-title>Spatial orientation assessment in the elderly: a comprehensive review of current tests</article-title><source>Brain Sci</source><year>2024</year><month>09</month><day>5</day><volume>14</volume><issue>9</issue><fpage>898</fpage><pub-id pub-id-type="doi">10.3390/brainsci14090898</pub-id><pub-id pub-id-type="medline">39335393</pub-id></nlm-citation></ref><ref id="ref43"><label>43</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Bartoschek</surname><given-names>T</given-names> </name><name name-style="western"><surname>Schwering</surname><given-names>A</given-names> </name><name name-style="western"><surname>Li</surname><given-names>R</given-names> </name><name name-style="western"><surname>M&#x00FC;nzer</surname><given-names>S</given-names> </name><name name-style="western"><surname>Carlos</surname><given-names>V</given-names> </name></person-group><person-group person-group-type="editor"><name name-style="western"><surname>Ahlqvist</surname><given-names>O</given-names> </name><name name-style="western"><surname>Schlieder</surname><given-names>C</given-names> </name></person-group><article-title>OriGami: a mobile geogame for spatial literacy</article-title><source>Geogames and Geoplay: Game-Based Approaches to the Analysis of Geo-Information</source><year>2018</year><publisher-name>Springer</publisher-name><fpage>37</fpage><lpage>62</lpage><pub-id pub-id-type="doi">10.1007/978-3-319-22774-0_3</pub-id></nlm-citation></ref><ref id="ref44"><label>44</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Schlieder</surname><given-names>C</given-names> </name><name name-style="western"><surname>Kiefer</surname><given-names>P</given-names> </name><name name-style="western"><surname>Matyas</surname><given-names>S</given-names> </name></person-group><article-title>Geogames: designing location-based games from classic board games</article-title><source>IEEE Intell Syst</source><year>2006</year><volume>21</volume><issue>5</issue><fpage>40</fpage><lpage>46</lpage><pub-id pub-id-type="doi">10.1109/MIS.2006.93</pub-id></nlm-citation></ref><ref id="ref45"><label>45</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>S&#x00E1;nchez-Escudero</surname><given-names>JP</given-names> </name><name name-style="western"><surname>Galvis-Herrera</surname><given-names>AM</given-names> </name><name name-style="western"><surname>S&#x00E1;nchez-Trujillo</surname><given-names>D</given-names> </name><etal/></person-group><article-title>Virtual reality and serious videogame-based instruments for assessing spatial navigation in Alzheimer's disease: a systematic review of psychometric properties</article-title><source>Neuropsychol Rev</source><year>2025</year><month>03</month><volume>35</volume><issue>1</issue><fpage>77</fpage><lpage>101</lpage><pub-id pub-id-type="doi">10.1007/s11065-024-09633-7</pub-id><pub-id pub-id-type="medline">38403731</pub-id></nlm-citation></ref><ref id="ref46"><label>46</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Wilczy&#x0144;ska</surname><given-names>D</given-names> </name><name name-style="western"><surname>Walczak-Koz&#x0142;owska</surname><given-names>T</given-names> </name><name name-style="western"><surname>Alarc&#x00F3;n</surname><given-names>D</given-names> </name><etal/></person-group><article-title>The role of immersive experience in anxiety reduction: evidence from virtual reality sessions</article-title><source>Brain Sci</source><year>2025</year><month>01</month><volume>15</volume><issue>1</issue><fpage>14</fpage><pub-id pub-id-type="doi">10.3390/brainsci15010014</pub-id></nlm-citation></ref><ref id="ref47"><label>47</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Zhu</surname><given-names>S</given-names> </name><name name-style="western"><surname>Lakshminarasimhan</surname><given-names>KJ</given-names> </name><name name-style="western"><surname>Arfaei</surname><given-names>N</given-names> </name><name name-style="western"><surname>Angelaki</surname><given-names>DE</given-names> </name></person-group><article-title>Eye movements reveal spatiotemporal dynamics of visually-informed planning in navigation</article-title><source>Elife</source><year>2022</year><month>05</month><day>3</day><volume>11</volume><fpage>e73097</fpage><pub-id pub-id-type="doi">10.7554/eLife.73097</pub-id><pub-id pub-id-type="medline">35503099</pub-id></nlm-citation></ref><ref id="ref48"><label>48</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Yi</surname><given-names>Z</given-names> </name><name name-style="western"><surname>Vaupel</surname><given-names>JW</given-names> </name></person-group><article-title>Functional capacity and self&#x2013;evaluation of health and life of oldest old in China</article-title><source>J Soc Issues</source><year>2002</year><month>01</month><volume>58</volume><issue>4</issue><fpage>733</fpage><lpage>748</lpage><pub-id pub-id-type="doi">10.1111/1540-4560.00287</pub-id></nlm-citation></ref><ref id="ref49"><label>49</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Duan</surname><given-names>J</given-names> </name><name name-style="western"><surname>Lv</surname><given-names>YB</given-names> </name><name name-style="western"><surname>Gao</surname><given-names>X</given-names> </name><etal/></person-group><article-title>Association of cognitive impairment and elderly mortality: differences between two cohorts ascertained 6-years apart in China</article-title><source>BMC Geriatr</source><year>2020</year><month>01</month><day>28</day><volume>20</volume><issue>1</issue><fpage>29</fpage><pub-id pub-id-type="doi">10.1186/s12877-020-1424-4</pub-id><pub-id pub-id-type="medline">31992221</pub-id></nlm-citation></ref><ref id="ref50"><label>50</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Nasreddine</surname><given-names>ZS</given-names> </name><name name-style="western"><surname>Phillips</surname><given-names>NA</given-names> </name><name name-style="western"><surname>B&#x00E9;dirian</surname><given-names>V</given-names> </name><etal/></person-group><article-title>The Montreal Cognitive Assessment, MoCA: a brief screening tool for mild cognitive impairment</article-title><source>J Am Geriatr Soc</source><year>2005</year><month>04</month><volume>53</volume><issue>4</issue><fpage>695</fpage><lpage>699</lpage><pub-id pub-id-type="doi">10.1111/j.1532-5415.2005.53221.x</pub-id></nlm-citation></ref><ref id="ref51"><label>51</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Li</surname><given-names>Z</given-names> </name><name name-style="western"><surname>Gong</surname><given-names>X</given-names> </name><name name-style="western"><surname>Wang</surname><given-names>S</given-names> </name><etal/></person-group><article-title>Cognitive impairment assessed by Mini-Mental State Examination predicts all-cause and CVD mortality in Chinese older adults: a 10-year follow-up study</article-title><source>Front Public Health</source><year>2022</year><volume>10</volume><fpage>908120</fpage><pub-id pub-id-type="doi">10.3389/fpubh.2022.908120</pub-id><pub-id pub-id-type="medline">36518570</pub-id></nlm-citation></ref><ref id="ref52"><label>52</label><nlm-citation citation-type="web"><article-title>Paper versions</article-title><source>MoCA Cognition</source><access-date>2026-03-13</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://mocacognition.com/paper/">https://mocacognition.com/paper/</ext-link></comment></nlm-citation></ref><ref id="ref53"><label>53</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Shin</surname><given-names>MS</given-names> </name><name name-style="western"><surname>Park</surname><given-names>SY</given-names> </name><name name-style="western"><surname>Park</surname><given-names>SR</given-names> </name><name name-style="western"><surname>Seol</surname><given-names>SH</given-names> </name><name name-style="western"><surname>Kwon</surname><given-names>JS</given-names> </name></person-group><article-title>Clinical and empirical applications of the Rey-Osterrieth Complex Figure Test</article-title><source>Nat Protoc</source><year>2006</year><volume>1</volume><issue>2</issue><fpage>892</fpage><lpage>899</lpage><pub-id pub-id-type="doi">10.1038/nprot.2006.115</pub-id><pub-id pub-id-type="medline">17406322</pub-id></nlm-citation></ref><ref id="ref54"><label>54</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Giovagnoli</surname><given-names>AR</given-names> </name><name name-style="western"><surname>Del Pesce</surname><given-names>M</given-names> </name><name name-style="western"><surname>Mascheroni</surname><given-names>S</given-names> </name><name name-style="western"><surname>Simoncelli</surname><given-names>M</given-names> </name><name name-style="western"><surname>Laiacona</surname><given-names>M</given-names> </name><name name-style="western"><surname>Capitani</surname><given-names>E</given-names> </name></person-group><article-title>Trail making test: normative values from 287 normal adult controls</article-title><source>Ital J Neurol Sci</source><year>1996</year><month>08</month><volume>17</volume><issue>4</issue><fpage>305</fpage><lpage>309</lpage><pub-id pub-id-type="doi">10.1007/BF01997792</pub-id><pub-id pub-id-type="medline">8915764</pub-id></nlm-citation></ref><ref id="ref55"><label>55</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Shepard</surname><given-names>RN</given-names> </name><name name-style="western"><surname>Metzler</surname><given-names>J</given-names> </name></person-group><article-title>Mental rotation of three-dimensional objects</article-title><source>Science</source><year>1971</year><month>02</month><day>19</day><volume>171</volume><issue>3972</issue><fpage>701</fpage><lpage>703</lpage><pub-id pub-id-type="doi">10.1126/science.171.3972.701</pub-id><pub-id pub-id-type="medline">5540314</pub-id></nlm-citation></ref><ref id="ref56"><label>56</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>H&#x00F6;lscher</surname><given-names>C</given-names> </name><name name-style="western"><surname>Br&#x00F6;samle</surname><given-names>M</given-names> </name><name name-style="western"><surname>Vrachliotis</surname><given-names>G</given-names> </name></person-group><article-title>Challenges in multilevel wayfinding: a case study with the space syntax technique</article-title><source>Environ Plann B Plann Des</source><year>2012</year><month>02</month><volume>39</volume><issue>1</issue><fpage>63</fpage><lpage>82</lpage><pub-id pub-id-type="doi">10.1068/b34050t</pub-id></nlm-citation></ref><ref id="ref57"><label>57</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Tom&#x00E9;</surname><given-names>A</given-names> </name><name name-style="western"><surname>Kuipers</surname><given-names>M</given-names> </name><name name-style="western"><surname>Pinheiro</surname><given-names>T</given-names> </name><name name-style="western"><surname>Nunes</surname><given-names>M</given-names> </name><name name-style="western"><surname>Heitor</surname><given-names>T</given-names> </name></person-group><article-title>Space&#x2013;use analysis through computer vision</article-title><source>Autom Constr</source><year>2015</year><month>09</month><volume>57</volume><fpage>80</fpage><lpage>97</lpage><pub-id pub-id-type="doi">10.1016/j.autcon.2015.04.013</pub-id></nlm-citation></ref><ref id="ref58"><label>58</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Hirtle</surname><given-names>SC</given-names> </name><name name-style="western"><surname>Jonides</surname><given-names>J</given-names> </name></person-group><article-title>Evidence of hierarchies in cognitive maps</article-title><source>Mem Cognit</source><year>1985</year><month>05</month><volume>13</volume><issue>3</issue><fpage>208</fpage><lpage>217</lpage><pub-id pub-id-type="doi">10.3758/bf03197683</pub-id><pub-id pub-id-type="medline">4046821</pub-id></nlm-citation></ref><ref id="ref59"><label>59</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Bendjedidi</surname><given-names>S</given-names> </name><name name-style="western"><surname>Bada</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Meziani</surname><given-names>R</given-names> </name></person-group><article-title>Urban plaza design process using space syntax analysis: El-Houria plaza, Biskra, Algeria</article-title><source>Int Rev Spat Plan Sustain Dev A Plan Strateg Des</source><year>2019</year><volume>7</volume><issue>2</issue><fpage>125</fpage><lpage>142</lpage><pub-id pub-id-type="doi">10.14246/irspsda.7.2_125</pub-id></nlm-citation></ref><ref id="ref60"><label>60</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Iftikhar</surname><given-names>H</given-names> </name><name name-style="western"><surname>Shah</surname><given-names>P</given-names> </name><name name-style="western"><surname>Luximon</surname><given-names>Y</given-names> </name></person-group><article-title>Human wayfinding behaviour and metrics in complex environments: a systematic literature review</article-title><source>Archit Sci Rev</source><year>2021</year><volume>64</volume><fpage>452</fpage><lpage>643</lpage><pub-id pub-id-type="doi">10.1080/00038628.2020.1777386</pub-id></nlm-citation></ref><ref id="ref61"><label>61</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Montello</surname><given-names>DR</given-names> </name></person-group><person-group person-group-type="editor"><name name-style="western"><surname>Shah</surname><given-names>P</given-names> </name><name name-style="western"><surname>Miyake</surname><given-names>A</given-names> </name></person-group><article-title>Navigation</article-title><source>The Cambridge Handbook of Visuospatial Thinking</source><year>2005</year><publisher-name>Cambridge University Press</publisher-name><fpage>257</fpage><lpage>294</lpage><pub-id pub-id-type="doi">10.1017/CBO9780511610448.008</pub-id></nlm-citation></ref><ref id="ref62"><label>62</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Xia</surname><given-names>J (Cecilia</given-names> </name><name name-style="western"><surname>Arrowsmith</surname><given-names>C</given-names> </name><name name-style="western"><surname>Jackson</surname><given-names>M</given-names> </name><name name-style="western"><surname>Cartwright</surname><given-names>W</given-names> </name></person-group><article-title>The wayfinding process relationships between decision-making and landmark utility</article-title><source>Tour Manag</source><year>2008</year><month>06</month><volume>29</volume><issue>3</issue><fpage>445</fpage><lpage>457</lpage><pub-id pub-id-type="doi">10.1016/j.tourman.2007.05.010</pub-id></nlm-citation></ref><ref id="ref63"><label>63</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Irmischer</surname><given-names>IJ</given-names> </name><name name-style="western"><surname>Clarke</surname><given-names>KC</given-names> </name></person-group><article-title>Measuring and modeling the speed of human navigation</article-title><source>Cartogr Geogr Inf Sci</source><year>2018</year><month>03</month><day>4</day><volume>45</volume><issue>2</issue><fpage>177</fpage><lpage>186</lpage><pub-id pub-id-type="doi">10.1080/15230406.2017.1292150</pub-id></nlm-citation></ref><ref id="ref64"><label>64</label><nlm-citation citation-type="report"><person-group person-group-type="author"><name name-style="western"><surname>MacQueen</surname><given-names>J</given-names> </name></person-group><article-title>Some methods for classification and analysis of multivariate observations</article-title><year>1967</year><access-date>2026-03-09</access-date><publisher-name>University of California Press</publisher-name><comment><ext-link ext-link-type="uri" xlink:href="https://scispace.com/pdf/some-methods-for-classification-and-analysis-of-multivariate-4pswti19oz.pdf">https://scispace.com/pdf/some-methods-for-classification-and-analysis-of-multivariate-4pswti19oz.pdf</ext-link></comment></nlm-citation></ref><ref id="ref65"><label>65</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Archdeacon</surname><given-names>TJ</given-names> </name></person-group><source>Correlation and Regression Analysis: A Historian&#x2019;s Guide</source><year>1994</year><access-date>2026-03-09</access-date><publisher-name>University of Wisconsin Press</publisher-name><comment><ext-link ext-link-type="uri" xlink:href="https://history.wisc.edu/publications/correlation-and-regression-analysis-a-historians-guide/">https://history.wisc.edu/publications/correlation-and-regression-analysis-a-historians-guide/</ext-link></comment><pub-id pub-id-type="other">9780299136543</pub-id></nlm-citation></ref><ref id="ref66"><label>66</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Thinsungnoen</surname><given-names>T</given-names> </name><name name-style="western"><surname>Kaoungku</surname><given-names>N</given-names> </name><name name-style="western"><surname>Durongdumronchai</surname><given-names>P</given-names> </name><name name-style="western"><surname>Kerdprasop</surname><given-names>K</given-names> </name><name name-style="western"><surname>Kerdprasop</surname><given-names>N</given-names> </name></person-group><person-group person-group-type="editor"><name name-style="western"><surname>Kerdprasop</surname><given-names>N</given-names> </name></person-group><article-title>The clustering validity with silhouette and sum of squared errors</article-title><year>2015</year><conf-name>International Conference on Industrial Application Engineering 2015</conf-name><conf-date>Mar 28-31, 2015</conf-date><pub-id pub-id-type="doi">10.12792/iciae2015.012</pub-id></nlm-citation></ref><ref id="ref67"><label>67</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Rousseeuw</surname><given-names>PJ</given-names> </name></person-group><article-title>Silhouettes: a graphical aid to the interpretation and validation of cluster analysis</article-title><source>J Comput Appl Math</source><year>1987</year><month>11</month><volume>20</volume><fpage>53</fpage><lpage>65</lpage><pub-id pub-id-type="doi">10.1016/0377-0427(87)90125-7</pub-id></nlm-citation></ref><ref id="ref68"><label>68</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Claessen</surname><given-names>MHG</given-names> </name><name name-style="western"><surname>van der Ham</surname><given-names>IJM</given-names> </name></person-group><article-title>Classification of navigation impairment: a systematic review of neuropsychological case studies</article-title><source>Neurosci Biobehav Rev</source><year>2017</year><month>02</month><volume>73</volume><fpage>81</fpage><lpage>97</lpage><pub-id pub-id-type="doi">10.1016/j.neubiorev.2016.12.015</pub-id><pub-id pub-id-type="medline">27993606</pub-id></nlm-citation></ref><ref id="ref69"><label>69</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Wiener</surname><given-names>JM</given-names> </name><name name-style="western"><surname>B&#x00FC;chner</surname><given-names>SJ</given-names> </name><name name-style="western"><surname>H&#x00F6;lscher</surname><given-names>C</given-names> </name></person-group><article-title>Taxonomy of human wayfinding tasks: a knowledge-based approach</article-title><source>Spat Cogn Comput</source><year>2009</year><month>05</month><day>20</day><volume>9</volume><issue>2</issue><fpage>152</fpage><lpage>165</lpage><pub-id pub-id-type="doi">10.1080/13875860902906496</pub-id></nlm-citation></ref><ref id="ref70"><label>70</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Cogn&#x00E9;</surname><given-names>M</given-names> </name><name name-style="western"><surname>Taillade</surname><given-names>M</given-names> </name><name name-style="western"><surname>N&#x2019;Kaoua</surname><given-names>B</given-names> </name><etal/></person-group><article-title>The contribution of virtual reality to the diagnosis of spatial navigation disorders and to the study of the role of navigational aids: a systematic literature review</article-title><source>Ann Phys Rehabil Med</source><year>2017</year><month>06</month><volume>60</volume><issue>3</issue><fpage>164</fpage><lpage>176</lpage><pub-id pub-id-type="doi">10.1016/j.rehab.2015.12.004</pub-id><pub-id pub-id-type="medline">27017533</pub-id></nlm-citation></ref><ref id="ref71"><label>71</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Wei</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Liu</surname><given-names>J</given-names> </name><name name-style="western"><surname>Jin</surname><given-names>L</given-names> </name><etal/></person-group><article-title>Individual behavior and attention distribution during wayfinding for emergency shelter: an eye-tracking study</article-title><source>Sustainability</source><year>2023</year><month>08</month><volume>15</volume><issue>15</issue><fpage>11880</fpage><pub-id pub-id-type="doi">10.3390/su151511880</pub-id></nlm-citation></ref><ref id="ref72"><label>72</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Kalantari</surname><given-names>S</given-names> </name><name name-style="western"><surname>Mostafavi</surname><given-names>A</given-names> </name><name name-style="western"><surname>Xu</surname><given-names>TB</given-names> </name><name name-style="western"><surname>Lee</surname><given-names>AS</given-names> </name><name name-style="western"><surname>Yang</surname><given-names>Q</given-names> </name></person-group><article-title>Comparing spatial navigation in a virtual environment vs. an identical real environment across the adult lifespan</article-title><source>Comput Human Behav</source><year>2024</year><month>08</month><volume>157</volume><fpage>108210</fpage><pub-id pub-id-type="doi">10.1016/j.chb.2024.108210</pub-id></nlm-citation></ref><ref id="ref73"><label>73</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Byrne</surname><given-names>P</given-names> </name><name name-style="western"><surname>Becker</surname><given-names>S</given-names> </name><name name-style="western"><surname>Burgess</surname><given-names>N</given-names> </name></person-group><article-title>Remembering the past and imagining the future: a neural model of spatial memory and imagery</article-title><source>Psychol Rev</source><year>2007</year><month>04</month><volume>114</volume><issue>2</issue><fpage>340</fpage><lpage>375</lpage><pub-id pub-id-type="doi">10.1037/0033-295X.114.2.340</pub-id><pub-id pub-id-type="medline">17500630</pub-id></nlm-citation></ref><ref id="ref74"><label>74</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Luo</surname><given-names>CM</given-names> </name><name name-style="western"><surname>Proctor</surname><given-names>RW</given-names> </name></person-group><article-title>How different direct association routes influence the indirect route in the same Simon-like task</article-title><source>Psychol Res</source><year>2019</year><month>11</month><volume>83</volume><issue>8</issue><fpage>1733</fpage><lpage>1748</lpage><pub-id pub-id-type="doi">10.1007/s00426-018-1024-5</pub-id></nlm-citation></ref><ref id="ref75"><label>75</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Jin</surname><given-names>WJ</given-names> </name><name name-style="western"><surname>Qin</surname><given-names>H</given-names> </name><name name-style="western"><surname>Zhang</surname><given-names>K</given-names> </name><name name-style="western"><surname>Chen</surname><given-names>XW</given-names> </name></person-group><person-group person-group-type="editor"><name name-style="western"><surname>Wang</surname><given-names>H</given-names> </name></person-group><article-title>Spatial navigation</article-title><source>Neural Circuits of Innate Behaviors</source><year>2020</year><publisher-name>Springer</publisher-name><fpage>63</fpage><lpage>90</lpage><pub-id pub-id-type="doi">10.1007/978-981-15-7086-5_7</pub-id></nlm-citation></ref><ref id="ref76"><label>76</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Savino</surname><given-names>GL</given-names> </name><name name-style="western"><surname>Emanuel</surname><given-names>N</given-names> </name><name name-style="western"><surname>Kowalzik</surname><given-names>S</given-names> </name><etal/></person-group><article-title>Comparing pedestrian navigation methods in virtual reality and real life</article-title><conf-name>ICMI &#x2019;19: 2019 International Conference on Multimodal Interaction</conf-name><conf-date>Oct 14-18, 2019</conf-date><pub-id pub-id-type="doi">10.1145/3340555.335374</pub-id></nlm-citation></ref><ref id="ref77"><label>77</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Hirsh</surname><given-names>JB</given-names> </name><name name-style="western"><surname>Mar</surname><given-names>RA</given-names> </name><name name-style="western"><surname>Peterson</surname><given-names>JB</given-names> </name></person-group><article-title>Psychological entropy: a framework for understanding uncertainty-related anxiety</article-title><source>Psychol Rev</source><year>2012</year><month>04</month><volume>119</volume><issue>2</issue><fpage>304</fpage><lpage>320</lpage><pub-id pub-id-type="doi">10.1037/a0026767</pub-id><pub-id pub-id-type="medline">22250757</pub-id></nlm-citation></ref><ref id="ref78"><label>78</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Lacreuse</surname><given-names>A</given-names> </name><name name-style="western"><surname>Raz</surname><given-names>N</given-names> </name><name name-style="western"><surname>Schmidtke</surname><given-names>D</given-names> </name><name name-style="western"><surname>Hopkins</surname><given-names>WD</given-names> </name><name name-style="western"><surname>Herndon</surname><given-names>JG</given-names> </name></person-group><article-title>Age-related decline in executive function as a hallmark of cognitive ageing in primates: an overview of cognitive and neurobiological studies</article-title><source>Philos Trans R Soc Lond B Biol Sci</source><year>2020</year><month>11</month><day>9</day><volume>375</volume><issue>1811</issue><fpage>20190618</fpage><pub-id pub-id-type="doi">10.1098/rstb.2019.0618</pub-id><pub-id pub-id-type="medline">32951543</pub-id></nlm-citation></ref><ref id="ref79"><label>79</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Xu</surname><given-names>TB</given-names> </name><name name-style="western"><surname>Govani</surname><given-names>VN</given-names> </name><name name-style="western"><surname>Kalantari</surname><given-names>S</given-names> </name></person-group><article-title>Age differences in pedestrian navigational skills and performance: a systemic review and meta-analysis</article-title><source>Ageing Res Rev</source><year>2025</year><month>02</month><volume>104</volume><fpage>102591</fpage><pub-id pub-id-type="doi">10.1016/j.arr.2024.102591</pub-id><pub-id pub-id-type="medline">39710072</pub-id></nlm-citation></ref><ref id="ref80"><label>80</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Ma</surname><given-names>Q</given-names> </name><name name-style="western"><surname>Wei</surname><given-names>X</given-names> </name><name name-style="western"><surname>Ren</surname><given-names>G</given-names> </name></person-group><article-title>Street design guidelines and urban road system improvement: from traffic volume to space quality</article-title><source>Transp Res Procedia</source><year>2025</year><volume>82</volume><fpage>3266</fpage><lpage>3282</lpage><pub-id pub-id-type="doi">10.1016/j.trpro.2024.12.095</pub-id></nlm-citation></ref><ref id="ref81"><label>81</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>van der Ham</surname><given-names>IJM</given-names> </name><name name-style="western"><surname>Faber</surname><given-names>AME</given-names> </name><name name-style="western"><surname>Venselaar</surname><given-names>M</given-names> </name><name name-style="western"><surname>van Kreveld</surname><given-names>MJ</given-names> </name><name name-style="western"><surname>L&#x00F6;ffler</surname><given-names>M</given-names> </name></person-group><article-title>Ecological validity of virtual environments to assess human navigation ability</article-title><source>Front Psychol</source><year>2015</year><volume>6</volume><issue>637</issue><fpage>637</fpage><pub-id pub-id-type="doi">10.3389/fpsyg.2015.00637</pub-id><pub-id pub-id-type="medline">26074831</pub-id></nlm-citation></ref></ref-list><app-group><supplementary-material id="app1"><label>Multimedia Appendix 1</label><p>Instructions presented on the desktop during each phase of the navigation game.</p><media xlink:href="games_v14i1e83128_app1.png" xlink:title="PNG File, 268 KB"/></supplementary-material><supplementary-material id="app2"><label>Multimedia Appendix 2</label><p>(A) and (B) show the sum of squared errors and silhouette coefficient, respectively, for younger and older adults across different values of k.</p><media xlink:href="games_v14i1e83128_app2.png" xlink:title="PNG File, 253 KB"/></supplementary-material><supplementary-material id="app3"><label>Multimedia Appendix 3</label><p>(A) and (B) show all instantaneous speed samples from older and young participants, respectively. Blue, green, and yellow points represent low-, medium-, and high-speed clusters, respectively, and black points indicate the cluster centroids. (C) and (D) present the probability distributions of navigation speed for older and younger participants, respectively, with dark blue representing the group mean and light blue indicating individual distributions.</p><media xlink:href="games_v14i1e83128_app3.png" xlink:title="PNG File, 2365 KB"/></supplementary-material></app-group></back></article>