Abstract
In this report, we present the project URWalking conducted at the University of Regensburg. We describe its major outcomes: Firstly, an indoor navigation system for pedestrians as a web application and as an Android app with position tracking of users in indoor and outdoor environments. Our implementation showcases that a variant of the \(A^*\)-algorithm by Ullmann (tengetriebene optimierung präferenzadaptiver fußwegrouten durch gebäudekomplexe https://epub.uni-regensburg.de/43697/, 2020) can handle the routing problem in large, levelled indoor environments efficiently. Secondly, the apps have been used in several studies for a deeper understanding of human wayfinding. We collected eye tracking and synchronized video data, think aloud protocols, and log data of users interacting with the apps. We applied state-of-the-art deep learning models for gaze tracking and automatic classification of landmarks. Our results indicate that even the most recent version of the YOLO image classifier by Redmon and Farhadi (olov3: An incremental improvement. arXiv, 2018) needs finetuning to recognize everyday objects in indoor environments. Furthermore, we provide empirical evidence that appropriate machine learning models are helpful to bridge behavioural data from users during wayfinding and conceptual models for the salience of objects and landmarks. However, simplistic models are insufficient to reasonably explain wayfinding behaviour in real time—an open issue in GeoAI. We conclude that the GeoAI community should collect more naturalistic log data of wayfinding activities in order to build efficient machine learning models capable of predicting user reactions to routing instructions and of explaining how humans integrate stimuli from the environment as essential information into routing instructions while solving wayfinding tasks. Such models form the basis for real-time wayfinding assistance.
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.Avoid common mistakes on your manuscript.
1 Overview of the Project
URWalking started as a student project for indoor navigation at the University of Regensburg. Later on, the project was integrated into the research project NADINE (funded by BMBFFootnote 1) to implement a navigation aid for public transport users for all public transport stops in Nuremberg (Germany). Furthermore, the system was incorporated into the research project DIVIS (funded by IuK Bayern) for the implementation of advanced indoor tracking strategies using mainly inertial sensors of smartphones combined with spatial and behavioural knowledge. We have developed a web application for usage in any browser and an smartphone app for Android devices. The interface was iterately improved by updating the implementation incorporating results of evaluations of previous versions [2, 19].
URWalking’s knowledge base connects 5,130 rooms in 27 buildings of the university itself and the nearby University of Applied Sciences. The area covered by the system exhibits the following quantitative features:
Space in total | 4,3 km\(^2\) |
Length of all paths | 110 km |
Doors | 1.563 |
Elevators | 320 |
Stairs | 151 |
We have also drawn maps for each floor in each of the modelled buildings. In this way, we have implemented results from several user studies [3, 17].
Currently, the system is in daily use at the University of Regensburg as a web application and an Android app. It processes between 100 and 5,000 route inquiries per day—a typical situation for a university navigation system: with a new term starting, many persons are new to the campus and seek assistance to find locations. The inquiries decrease as soon as these persons have learned a spatial map of the campus.
2 Components of the System
URWalking consists of four main components: the basic component is the web server that stores all maps of the covered environment and calculates routes. For URWalking to be used in a browser, we developed a web client as the second component of the URWalking system (see Fig. 1). As the web client lacks indoor positioning, users can give interactive feedback when they want to see the next routing instruction. The third component is our Android application. It provides the same functionality as the web client plus indoor tracking of the user’s current position along the route (see Fig. 2). The fourth component is YAMA—a web app for creating, editing, and maintaining map data (see Fig. 4).
2.1 URWalking Web Server
Routes are calculated by a shortest path algorithm. Edge weights have been learned using genetic optimisation in order to reflect human route preferences that often lead to routes that are very distinct from shortest routes in terms of time or distance [15, 24, 25]. When choosing a destination, users can choose which type of route calculation they prefer. Routes that mimic human wayfinding decisions are typically longer than shortest routes, but less complex in terms of transitions between indoor and outdoor and changes in walking directions (see Fig. 3). These preferences seem linked to a person’s spatial map of an environment and the person’s capability to recall details. It is an challenging issue for future research to analyse this relationship in detail.
The web server provides an API used by the both client applications for route calculation and access to maps and route instructions (see Figs. 1 and 2).
2.2 Web Application
The purpose of the web app is to visualize maps and routing instructions in a browser window. As the app runs on all common browsers, it requires no installation procedure and is used by the majority of the users.
2.3 Android Application
The Android application (see Fig. 2) relies on the web server for calcuations of routes and appropriate instructions. It is available on Google Play StoreFootnote 2 and constitutes our experimental framework for adding indoor tracking of users in order to provide a user experience similar to car navigation systems and update maps and routing instructions automatically. In February 2022, there were around 1,700 active installations. Reliable indoor tracking of users for several hours they spend on the campus performing very different activities is still an open research issue.
2.4 Data for Indoor Navigation
For calculating routes and generating routing instructions, the web server uses maps that model the environment [24]. Our concept for mapping indoor environments is hybrid. Firstly, it is graph-based (for accessibility relations between locations that are needed for route calculation). Secondly, it is hierarchical in order to formalize the structure of the environment (separate areas of an environment, separate buildings in an area, separate levels in each building). Thirdly, it is also semantic: nodes come in several categories in order to capture various functions of locations in indoor buildings (e.g. doors, toilettes, offices, corridors, or landmarks—see Fig. 4). Edges can also be of different category: e.g. indoor connections between two nodes, outdoor connections, stairs, elevators, street crossings. Node and edge categories allow route calculation to account for user preferences [15]: e.g. handicapped users prefer elevators over stairs. Finally, it is spatial: as for indoor position tracking we snap positions to edges as the most probable footways on a route, sometimes we have to consider spatial dimensions are locations: e.g. for broad corridors a grid is a better model to allow higher degrees of freedom for users to move in the corridor.
For visualizing the model as maps of indoor areas, we used Inkscape to generating scalable maps in the SVG vector graphics format. All other (symbolic) data can be created, edited, and maintained with the editing tool YAMA (see Fig. 4) that is part of URWalking.
As all data for running URWalking in an environment can be configured with YAMA, URWalking can be configured for any environment repeating our systematic procedure to create data for the navigation service on the campus. For this purpose, our software is available as freeware upon requestFootnote 3.
3 Applications in Research Contexts
URWalking implements path planning based on state-of-the-art algorithms. It showcases that real-time route calculation is tractable in a client server framework with several thousand requests per day, many of them concurrently. For routing instructions, URWalking according to recent findings in the literature [1, 5, 12, 26] generates instructions incorporating the most salient landmark close to the user. The instructions are generated completely automatically by inspecting their properties for being visible in advance [13, 22]. To the best of our knowledge, no other freely available pedestrian indoor navigation system offers these features.
A further unique property of URWalking is that it runs since several years as a service for members of the university. As we log usage data, we are collecting a corpus of naturalistic user data in realistic settings. Furthermore, we conduct experiments using URWalking to investigate two important research questions in assisted wayfinding: (1) Can we predict how users perceive the current situation at any time during a wayfinding task in real-time and what decision they will take next? (2) Can we validate the role of conceptual models claiming to contribute to an answer by building machine learning models implementing these concepts?
3.1 Data Driven Validation of Wayfinding Models
A prominent conceptual model of human decision making during wayfinding such as that proposed in [8]. The authors measure the complexity of a routing instruction (type) \(t_i\) in an outdoor environment e:
Here, \(\text{ b }(t_i,e)\) is the branching factor, i.e. the number of options for the decision how to continue a route. \(\text{ r }(t_i,e)\) is the ease of detecting a mentioned landmark in the physical environment, and \(\text{ v }(t_i,e)\) is the visibility in advance of landmarks. In order to understand how such models can be validated, adjusted, or modified for indoor environments, we conducted various controlled eye tracking studies. These are controlled experiments in which we record video data synchronized with the gaze data (obtained from the SMI Eye Tracking Glasses 2) and performance data such as hesitations or errors during wayfinding and the time needed to follow URWalking’s routing instructions.
In order to address question (1), we interpret \(\text{ b }(t_i,e)\) and \(\text{ r }(t_i,e)\) as predictors for how users will perceive a routing instruction at their current position. In our work, we try to find signals that allow use to implement a model that can serve as a proxy for \(\text{ b }(t_i,e)\) and \(\text{ r }(t_i,e)\), respectively.
In our analyses [1, p. 96], we found that—different to outdoor environments—\(\text{ b }(t_i,e)\) does not influence the task performance, i.e. correct human decisions. We conclude that indoor contexts seem to be clearer in terms of changes in direction due to architectural elements such as corridors, stairs, entries, or elevators.
For a data-driven operationalization of \(\text{ r }(t_i,e)\), we applied state-of-the-art machine learning models from computer vision to predict the visual salience of landmarks in route instructions from photographs of the landmarks [4]. We fine-tuned a pretrained VGG 19 CNN on the photographs with the respective salience as target variable. Results indicate that high level style, high level content, and visual complexity of the photographs are the best features the CNN can generate for landmark prediction. We conclude that \(\text{ r }(t_i,e)\) is correlated with \(\text{ v }(t_i,e)\). This observation would allow to automatically extract landmarks from visualisations of objects in indoor environments. As a consequence, we can simplify the the approach in [18]: There is no need to classify objects as POI to compute their salience. Instead, we can estimate it from photographs of the objects.
However, the limitation of this approach is that the photographs isolate landmarks and do not show them in their usual surroundings with many visual detractors.
3.2 Real-Time Prediction of Areas of Interest
In order to overcome this limitation, we decided to automatically recognize objects in the video stream gained from the Eye Tracking Glasses and thereby detect which objects users focus on. This could result in a better proxy for \(\text{ r }(t_i,e)\). For object recognition, we we used the YOLO [23] model as a state-of-the-art neural network. As our corpus was small, we used YOLO pretrained on the COCO dataset [16] without any fine-tuning. Our asumption was the COCO contained classes of objects that are typical in our indoor wayfinding video streams such as doors, stairs, or hallways (see Fig. 5).
The results in Tab. 1 indicate that the pretrained model recognizes too many classes with low confidence and accuracy values. From these results, we learned that fine-tuning is indispensable. For GeoAI, more specific data sets than COCO for indoor objects would be beneficial to automatically detect focussed objects during wayfinding. Such data would allow us to better understand which environmental stimuli influence human decision making. In order to contribute to the research question (2), we are currently annotating our data in order to finetune YOLO on the environment of the Regensburg campus. Our objective is to come up with an improved proxy for \(\text{ r }(t_i,e)\).
3.3 Prediction of Landmark Salience from Gaze Data
The lack of reliable object recognition being a drawback for finding a proxy for \(\text{ r }(t_i,e)\), we tried a less supervised appraoch: ignoring all visual data, we analysed the gaze data recorded sychronously to the video streams. We based our analysis on fixations. In earlier research, the fixation duration was used as an indicator for the difficulty in extracting the information processed [11], and the fixation frequency was considered as a factor of search efficiency [9]. These variables can be used to analyze the cognitive processes during wayfinding [14] and indicate how the next situation is perceived. Related results in eye tracking research point out the distinction between ambient and focal visual processing of visually perceived information [20]: During ambient processing information is explored superficially and input from peripheral vision may control eye movements. During focal processing central vision becomes dominant, the collected information is processed, and salient objects are recognized and interpreted.
As automated detection of fixations on the smartphone’s display was not reliable, we annotated each frame manually with a binary label: Is the gaze position on the navigation aid’s display or outside of it? From the annotated data, we could extract the frequency of fixations anywhere on the navigation aid’s display and outside of it (see Fig. 6).
To compute the frequencies of interest (on the display: (\(F_{\text{ on }}\), outside of the display: \(F_{\text{ out }}\)), we applied Empirical Mode Decomposition [7] on the gaze data: According to the results in e.g. [6, 20] focal processing is characterized by high fixation frequencies. As we worked with two different stimuli visible at the same time (display and environment), we normalized \(F_{\text{ out }}\) by \(F_{\text{ on }}\) and calculated a relation between the degree of focal processing outside and on the display. For that purpose, we defined the quotient PS (perceived salience of landmarks in routing instructions):
To avoid the case that PS may be undefined, we set \(PS=F_{\text{ out }}\) if \(F_{\text{ on }}=0\). Then, the intuition behind PS is the following: We can distinguish two cases:
-
\(PS>1\): the degree of focal processing in the environment is higher than that on the display. For the current routing instruction, the test person can focus his/her attention on few objects in the environment and does not need much effort to explore the environment to find the landmark referred to in the instruction.
-
\(PS\le 1\): the degree to focal processing in the environment is low in relation to that on the display. So, proportionally, test persons need more ambient processing to explore the environment in order to eventually locate the referred landmark.
Figure 7 visualizes the average (over all test persons) perceived salience score PS. The plot shows that despite of careful selection of the landmarks, PS is far from constant. We believe to have learned the following lesson: While the concept of PS appears quite simplistic, we can provide evidence, that for our data it correlates significantly with the concept of visual salience [12, 26] based on subjective self-reports (Spearman \(r = 0.656\), \(p = 0.0042\)). So, in fact, landmarks rated high for visual salience, are also perceived as visually salient in the complex physical environment in which they are embedded. On the one hand, this observation is exciting, as it allows PS to be interpreted as a real-time proxy for visual salience ratings. As a consequence, in the sense of research question (2) PS can be beneficial for automated generation of routing instructions referring to landmarks that are salient for the user at the moment of the instruction being given. This is a major advantage over ratings that are collected using questionnaires in a non-naturalistic way.
However, PS is to simplistic to explain the viewing process completely. GeoAI in the future should take the challenge to get more out of eye tracking data by applying more detailed models of viewing behaviour in order to understand the influence of architectural constraints imposed by the indoor surroundings, e.g. width of corridors. A second limitation of our study is the lack of information about which objects were fixated by test persons. An automated procedure to extract objects in the environment that were fixated significantly more often than others would be a great step towards automated identification of salient objects from gaze data. In this way, the bias of experimenters to choose areas of interest could be removed. This is another strong argument for GeoAI to create indoor wayfinding databases in order to finetune state-of-the-art image classifiers (see Sect. 3.2).
3.4 Real-Time Prediction of Assistance Needs
We can get, however, an idea of the influence indoor environments have on eye movements by further analysing fixations. As stated in Sect. 3.3, the PS score showed much more variance than the visual salience ratings of the landmarks chosen for the routing instructions. For a deeper analysis of this behaviour, we aggregated fixations between two routing instructions and generated heatmaps for all aggregations. We then used the distribution of dwell time [10, p. 535] for each aggregated heatmap as a measure for similarity of the gaze behaviour of our test persons. So we could also take the spatial distribution of fixations into account, not only their frequency anywhere in the environment.
With the landmarks in the routing instructions carefully chosen as the objects rated best between two subsequent wayfinding decision points and satisfying established criteria for landmarks [22] better than other objects, we assumed that the viewing behaviour should follow a similar pattern for each instruction: read it, identify it in the environment, and continue walking. Consequently, in two subsequent route segments the viewing behaviour should be similar if the environment did not have high impact on the gaze behaviour (e.g. by enforcing the user to take a turn at a crossing or while climbing a staircase). We quantified this impact by calculating the normalized mean square error between the observed dwell time distribution and its estimation from the distribution of the preceding aggregation. The resulting NMSE values for our test route are presented in Fig. 8. From an inspection of the segments with high and low NMSE, we conclude that the NMSE is low if there is no change in direction from a segment and the subsequent one, while it otherwise tends to increase. The highest values in segments 16, 17, and 20 have been calculated in staircases where persons have to change the direction while going up the stairs and reorient themselves continuously.
4 Conclusions and Current Research Interests
In this paper, we presented the URWalking system assisting users during indoor wayfinding. Usage data indicates that in real-time situations indoor navigation is appreciated by users although indoor positioning is not available as in our web application. From think aloud protocols we even know that no positioning is better than a system with poor performance in this aspect. In this sense, URWalking is innovative as most other indoor navigation systems try to solve the positioning issue first. URWalking, instead, first serves to collect large data sets that we will leverage to improve indoor positioning algorithms in the future.
Another important issue that still waits for a better solution is how to make URWalking better understand natural language descriptions of users for their destinations. Often, they do not know room identifiers. Our current implementation is capable of repairing minor errors in spelling; however, in many cases users describe an event they want to attend, a person that they want to meet, or a service of the university that they want to use. As we cannot control user inputs and do not have a redundant mechanism available for determining destinations, for most of the inquiries we lack reliable ground truth that we could use in machine learning approaches to improve the prediction accuracy of destinations users want to request wayfinding information for.
Finally, from the analyses discussed in this project report we learn that real-time tracking of human wayfinding behaviour is a difficult task and still needs progress in the GeoAI community on the construction of conceptual models, their empirical evaluation, and on AI algorithms for analysing wayfinding behaviour at run-time in order to provide situation specific assistance to users.
In order to move ahead, we currently try to find proxies for gaze data that we can only collect in controlled experiments, but not from users under naturalistic conditions. Therefore, it is one of the important issues on our research agenda to find out which interaction data can serve as proxies for human viewing behaviour. Beyond developing our models using data that we collect from URWalking users, we contribute to this field by sharing our log data with the communityFootnote 4, by integrating available indoor tracking implementations in our application, and by comparing their performance. In this way, we hope that our system can inspire the community to address many of issues that require solutions in order to develop better AI based wayfinding aids.
Notes
Please send an email to the corresponding author.
We are currently preparing a website to make the data available. Meanwhile, anybody interested in the data is invited to send an email to the corresponding author.
References
Bauer C (2018) Unterstützung der Orientierung im Innenbereich: Analyse landmarkenbasierter Karten-Interfaces anhand des Blickverhaltens der Nutzer. https://epub.uni-regensburg.de/37666/
Bauer C, Müller M, Ludwig B (2016) Indoor pedestrian navigation systems: Is more than one landmark needed for efficient self-localization? In: Proceedings of the 15th International Conference on Mobile and Ubiquitous Multimedia, MUM ’16, p. 75-79. Association for Computing Machinery, New York, NY, USA https://doi.org/10.1145/3012709.3012728
Bienk S, Kattenbeck M, Ludwig B, Müller M, Ohm C (2013) I want to view it my way: interfaces to mobile maps should adapt to the user’s orientation skills. In: Proceedings of the 12th International Conference on Mobile and Ubiquitous Multimedia, MUM ’13. Association for Computing Machinery, New York, NY, USA. https://doi.org/10.1145/2541831.2541841
Donabauer G, Ludwig B (2021) Testing landmark salience prediction in indoor environments based on visual information. GIScience. https://doi.org/10.25436/E2SG62
Duckham M, Winter S, Robinson M (2010) Including landmarks in routing instructions. J Locat Based Serv 4(1):28–52. https://doi.org/10.1080/17489721003785602
Eisenberg ML, Zacks JM (2016) Ambient and focal visual processing of naturalistic activity. J Vis 16(2):5–5
Flandrin P, Rilling G, Goncalves P (2004) Empirical mode decomposition as a filter bank. IEEE Signal Process Lett 11(2):112–114. https://doi.org/10.1109/LSP.2003.821662
Giannopoulos I, Kiefer P, Raubal M, Richter KF, Thrash T (2014) Wayfinding decision situations: a conceptual model and evaluation. In: Duckham M, Pebesma E, Stewart K, Frank AU (eds) Geographic information science. Springer International Publishing, Cham, pp 221–234
Goldberg JH, Kotval XP (1999) Computer interface evaluation using eye movements: methods and constructs. Int J Ind Ergon 24(6):631–645
Holmquvist K, Andersson R (2017) Eye tracking: a comprehensive guide to methods, paradigms, and measures. CreateSpace Independent Publishing Platform
Just MA, Carpenter P (1976) A eye fixations and cognitive processes. Cogn Psychol 8(4):441–480
Kattenbeck M (2015) Empirically measuring salience of objects for use in pedestrian navigation. In: Proceedings of the 23rd SIGSPATIAL International Conference on Advances in Geographic Information Systems, SIGSPATIAL ’15, pp. 3:1–3:10. ACM, New York, NY, USA
Kattenbeck M (2017) How subdimensions of salience influence each other. Comparing models based on empirical data. In: Clementini E, Donnelly M, Yuan M, Kray C, Fogliaroni P, Ballatore A (eds) 13th International Conference on Spatial Information Theory (COSIT 2017), Leibniz International Proceedings in Informatics (LIPIcs), vol. 86. Schloss Dagstuhl–Leibniz-Zentrum fuer Informatik, Dagstuhl, Germany, pp 10:1–10:13. https://doi.org/10.4230/LIPIcs.COSIT.2017.10
Kiefer P, Giannopoulos I, Raubal M (2014) Where am i? Investigating map matching during self-localization with mobile eye tracking in an urban environment. Trans GIS 18(5):660–686
Kreller I, Ludwig B (2021) User preferences and the shortest path. CoRR. https://arxiv.org/abs/2107.11150
Lin TY, Maire M, Belongie S, Bourdev L, Girshick R, Hays J, Perona P, Ramanan D, Zitnick CL, Dollár, P (2014) Microsoft coco: common objects in context. https://doi.org/10.48550/ARXIV.1405.0312
Ludwig B, Müller M, Ohm C (2014) Empirical evidence for context-aware interfaces to pedestrian navigation systems. KI - Künstliche Intelligenz 28(4):271–281. https://doi.org/10.1007/s13218-014-0333-0
Lyu H, Yu Z, Meng L (2015) A computational method for indoor landmark extraction. Springer International Publishing, Cham, pp 45–59. https://doi.org/10.1007/978-3-319-11879-6_4
Ohm C, Bienk S, Kattenbeck M, Ludwig B, Müller M (2016) Towards interfaces of mobile pedestrian navigation systems adapted to the user’s orientation skills. Pervasive and Mobile Computing, Thirteenth International Conference on Pervasive Computing and Communications (PerCom 2015), vol 26, pp 121–134
Pannasch S, Schulz J, Velichkovsky BM (2011) On the control of visual fixation durations in free viewing of complex images. Attention Percep Psychophys 73(4):1120–1132
Ramsauer, D (2019) Object detection for human gaze: Using deep learning for capturing areas of focus. Master’s thesis, University of Regensburg
Raubal M, Winter S (2002) Enriching wayfinding instructions with local landmarks. In: Egenhofer MJ, Mark DM (eds) Geographic information science. Springer, Berlin, pp 243–259
Redmon J, Farhadi A (2018) Yolov3: An incremental improvement. arxiv:1804.02767
Ullmann M (2020) Datengetriebene Optimierung präferenzadaptiver Fußwegrouten durch Gebäudekomplexe. https://epub.uni-regensburg.de/43697/
Ullmann M, Bauer C, Jackermeier R, Ludwig B (2018) Ain’t short sweet, when walking down that street? In: Adjunct proceedings of the 14th international conference on location based services, pp. 223–224. https://doi.org/10.3929/ethz-b-000225624
Winter S, Raubal M, Nothegger C (2005) Focalizing measures of salience for wayfinding. In: Meng L, Reichenbacher T, Zipf A (eds) Map-based mobile services. Springer, Berlin, Heidelberg. https://doi.org/10.1007/3-540-26982-7_9
Funding
Open Access funding enabled and organized by Projekt DEAL.
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Ludwig, B., Donabauer, G., Ramsauer, D. et al. URWalking: Indoor Navigation for Research and Daily Use. Künstl Intell 37, 83–90 (2023). https://doi.org/10.1007/s13218-022-00795-1
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s13218-022-00795-1