Abstract
This paper examines potential interaction aspects related to querying and the display of information in exploratory search scenarios with a particular focus on user state and interactive visualization. Exploratory search refers to a specific type of information seeking that is open-ended, continuous and evolving. The evolving nature of exploratory search also provides the computer with sequential data that can be used to estimate user state and intention as the search unfolds. In this setting, the system supports querying by relying on user’s pointing actions, sequential organization of user interaction and query metadata. The system also adapts the display of information by determining the timing and visual representation. The paper illustrates potential interactions that employ new input modalities such as eye gaze and physiological signals. The paper concludes by discussing the possible functions of interactive visualization regarding querying and the display of information.
Access provided by Autonomous University of Puebla. Download conference paper PDF
Similar content being viewed by others
Keywords
- Interaction design
- Physiological input
- Gaze input
- Interactive visualization
- Search context
- Exploratory search
1 Introduction
Digital/web search has become the primary tool for most of the information seeking tasks. Most search activity can be classified as basic look-up, searches that are known-item or fact retrieval tasks. Yet, long term search activities, such as exploring new domains of knowledge, are often more complex. Information seeking in such long term exploratory search often involves evaluation, comparison and synthesis of results and iterative querying. However, as expressed by Marchionini such open ended information needs are not well addressed in today’s search engines that are oriented towards high precision rather than maximizing the number of possibly relevant objects [1]. Related to this observation is the questioning of the dominant interaction model for search tools, “query-response paradigm”, and the reevaluation of guidelines for the design of systems to support exploratory search [2].
Typical query-response pair illustrates some common interactive qualities of search interfaces. Consider this sequence of events that occur during a basic look-up task:
A user types a query to a predefined search box in the graphical user interface and submits her query. The system responds by returning a set of results that are displayed as a list, sorted by their relevance as computed by the search engine.
The sequence reveals many aspects of interaction that include but are not limited to (a) the input modality (typing), (b) the initiation mechanism (explicitly initiated by user), and (c) the timing (concurrent), (d) presentation (textual) and (e) layout (list) of search results. In HCI various prototypes and design frameworks emerged that took a different approach to one or more of the various aspects listed above. This paper aims to identify future interactions with a particular focus on information visualization and various input modalities (such as EEG, EDR, fEMG, eye gaze and pupillometry which are currently positioned as peripheral in respect to more established and precise input techniques). Academic search, in which open ended and evolving searches is common, is used as a case.
Input modalities such as gaze and other physiological signals expand the resources available to the system for inferring the state of user [3], increasing the capability of the system for collaboration [4]. In search scenarios such input allows an increased ability of the computer in assigning user responses to information items. The increased awareness of user response also implies possible changes regarding how the user performs queries and how the system presents information in future search interfaces.
2 Querying
A potential outcome of human computer symbiosis in exploratory search scenarios is the possibility of performing search without having to articulate precise queries. In this setting, the system carries out the necessary work of formulating the query by possibly relying on user’s spatial references (pointing), the sequential organization of user actions and query metadata.
Pointing. Pointing, as a spatial referencing technique, allows directing attention to objects by using contextual information and is widely used in graphical interfaces through cursor and touch input. Pointing is also a possible method for facilitating the easy construction of queries as opposed to using typed queries. In most search interfaces pointing is used for selecting specific items, filters or links. Apart from these, several visual interfaces allow pointing to objects such as keywords or document surrogates for open ended querying. IntentRadar [5] features radially organized keywords as a representation of the user query and enables dragging the keywords for directing the search. Apolo [6] enables pointing to surrogates of scientific publications for query, with the system populating the visual interface with similar publications in response. In addition to pointing single items, pointing to multiple items or regions is possible by using interactive visualization and query relaxation techniques [7]. In common, these examples allow the user to form queries by using items that are readily available in the interface and assign the task of formulating the query to the system. On the other hand, possible forms of pointing are not limited to mouse or touch input. Spatial references can take many forms such as different hand postures, gaze and orientation of head, body and arms. Future interfaces can build on these different forms of indications for diversifying the possible modes of explicit inputs and for tracking user’s attention.
Sequential organization. Interaction during information seeking involves consecutive user actions while interacting with the user interface. The sequence of these actions constitutes part of what is called the search context and have been used to infer user preferences and to expand or disambiguate queries. Traditionally, most implicit mining of user data involved actions that are common to graphical user interfaces such as scrolling, selecting, navigating, saving, deleting [8], with few examples using alternative inputs such as gaze [9]. Incorporation of peripheral physiological input, allows more detailed information related to user state and interest level during the interaction, without having to rely on user’s explicit feedback. At the same time, the sequential organization does not always imply relation. Thus, one possible research area would be to identify gaps and adjacency pairs during interaction, or to provide the user the necessary means to designate search sequences.
Query metadata. In addition to sequential organization, the queries can be contextualized by using inputs that are concurrent to the primary input (Fig. 1). These can include both the metadata of the primary input (e.g. the pace of typing when entering a query), or peripheral input that accompany the primary input (e.g. accompanying eye gaze or physiological signals while the user is typing a query). These metadata can be used to clarify user’s query, similar to the role of intonation, pitch and amplitude for interpreting participants’ utterances in daily conversation [10]. An HCI example of using input metadata is registering pressure information while typing on a small keyboard to explicitly control the level of input uncertainty [11].
3 Display of Information
Another outcome of human computer symbiosis is adapting the display of information to the user state by relying on detailed information about user’s attention level and response. The system utilizes this information to determine the timing and visual representation of information items.
Timing. In typical search interfaces the display of the results is concurrent: all the results are displayed at once. In contrast, future systems can use time order expressively to highlight and relate certain items within the retrieved result set by the order of their introduction. Gaze input in these cases can be used to detect user’s attention levels, in turn affecting the pace of display.
Besides using timing in response to a query, a radical alternative is to eliminate queries and proactively retrieve information based on user’s changing context. In HCI various design frameworks exist such as non-command [12], mixed-initiative [13] and attentive user interfaces [14] that promote information retrieval without users having to explicitly query, but rather rely on software agents [15]. Such proactive display of information, however, potentially causes the problem of distracting the user from the task at hand. As Allen [16] has noted, as opposed to fixed initiative systems in which the initiation of interaction is well-defined, in mixed initiative systems the agents should decide on the appropriate time of starting the interaction. User actions, as well as gaze and other physiological signals in this case can potentially provide necessary triggers for the timing of information display.
Visual representation. Search interfaces usually represent retrieved set of results in surrogates, namely the representation of original documents. Surrogates can also indicate various user actions performed by the user. Traditionally these indications relied on the explicit input of the user. Typical examples are the change of color for visited Web pages in a search result list and the display of actions such as forwarding, replying and reading for emails. EEG and other physiological inputs provide the opportunity to visually distinguish information items without any explicit input (Fig. 2), by tracking user attention (whether the surrogate is gazed at by the user) and physiological response (how the interaction with the document affected the cognitive load and arousal level of the user). Interaction history can also enable tailoring the layout of surrogates in the visual space by taking the user interaction history into account. A possible example is changing the position of a document in a graph after user interaction, by registering what the user specifically engaged within the document. In this scenario possible clues for the change include sequential organization of search as well as various physiological inputs recorded while the user is engaged with the different parts of the document.
4 Discussion: Roles of Visualization
This section identifies possible roles for the visualization of items during search regarding the querying and display of information examined above.
As a resource for pointing. Visual display of information provides set of items that the user can point to for explicit querying. In addition to querying, visualization enables contextualizing the search through sequential organization or query metadata by pointing to the items prior to or during the query. These pointing actions can be used to indicate not only the items themselves but the specific user interaction history with the item (e.g. what the user has specifically found relevant within the item). In this case the registration of user’s previous response allows referring to the past context of interaction.
Orient the user. Search results from various stages of a search session can be visualized to orient the user by juxtaposing newly retrieved information items with familiar ones. Visual representation of the previously interacted items, in this case, indicates how familiar the user is with an item and her past implicit response during interaction.
Prioritize events and information items. Visualization enables prioritizing information items and events by displaying them in different visual areas and with different visual features. Together with timing, visualization enables notifying the user of possible actions that are initiated by the system such as adding, removing or highlighting items.
Represent system estimation of user intention. The systems estimation of user state or intention can be conveyed through visual representation, for the user to make sense of system actions and repair any miscommunications.
5 Conclusion and Future Work
The paper identified potential design directions regarding querying and display of information in future search scenarios by focusing on new input modalities and information visualization. These design directions also described how human computer symbiosis can be furthered with the increasing role of computers for formulating queries and adapting the display to the user. The review of the above interactive aspects is by no means exhaustive but is intended to spark a discussion for future design. Future work will focus on prototyping these different dimensions for realistic search scenarios.
References
Marchionini, G.: Exploratory search: from finding to understanding. Commun. ACM 49(4), 41–46 (2006)
White, R.W., Roth, R.A.: Exploratory search: beyond the query-response paradigm. In: Synthesis Lectures on Information Concepts, Retrieval, and Services. Morgan & Claypool (2009)
Barral O., Jacucci G.: Applying physiological computing methods to study psychological, affective and motivational relevance. In: Proceedings of Symbiotic’14. Springer Berlin Heidelberg (2014)
Jacucci, G., Spagnolli, A., Freeman, J., Gamberini, L.: Symbiotic interaction: a critical definition and comparison to other human-computer paradigms. In: Proceedings of Symbiotic’14. Springer Berlin Heidelberg (2014)
Ruotsalo, T., Peltonen, J., Eugster, M., Głowacka, D., Konyushkova, K., Athukorala, K., Jacucci, G., Kaski, S.: Directing exploratory search with interactive intent modeling. In: Proceedings of CIKM’13, 1759-1764. ACM Press, New York (2013)
Chau, D.H., Kittur, A., Hong, J.I., Faloutsos, C.: Apolo: making sense of large network data by combining rich user interaction and machine learning. In: Proceedings of CHI’13, pp. 167–176. ACM Press, New York (2013)
Heer, J., Agrawala, M., Willett, W.: Generalized selection via interactive query relaxation. In: Proceedings of CHI’08, pp. 959–968. ACM Press, New York (2008)
Kelly, D., Teevan, J.: Implicit feedback for inferring user preference: a bibliography. In: ACM SIGIR Forum, vol. 37, issue 2, pp. 18–28. ACM Press, New York (2003)
Maglio, P.P., Barrett, R., Campbell, C.S., Selker, T.: SUITOR: An attentive information system. In: Proceedings of IUI’00, pp. 169–176. ACM Press, New York (2000)
Tannen, D.: Conversational style. Ablex, Norwood (1984)
Weir, D., Pohl, H., Rogers, S., Vertanen, K., Kristensson, P.O.: Uncertain text entry on mobile devices. In: Proceedings of CHI’14, pp. 2307–2316. ACM Press, New York (2014)
Nielsen, J.: Noncommand user interfaces. Commun. ACM 36(4), 83–99 (1993). ACM Press, New York
Horvitz, E.: Principles of mixed-initiative user interfaces. In: Proceedings of CHI’99, pp. 159–166. ACM Press, New York (1999)
McCrickard, D.S., Chewar, C.M.: Attuning notification design to user goals and attention costs. Commun. ACM 46(3), 67–72 (2003). ACM Press, New York
Rhodes, B.J., Maes, P.: Just-in-time information retrieval agents. IBM Syst. J. 39(4), 685–704 (2000)
Allen, J.E., Guinn, C.I., Horvitz, E.: Mixed-initiative interaction. IEEE Intell. Syst. Appl. 14(5), 14–23 (1999)
Acknowledgements
This work has been partly supported by MindSee (FP7 – ICT; Grant Agreement # 611570).
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2014 Springer International Publishing Switzerland
About this paper
Cite this paper
Serim, B. (2014). Querying and Display of Information: Symbiosis in Exploratory Search Interaction Scenarios. In: Jacucci, G., Gamberini, L., Freeman, J., Spagnolli, A. (eds) Symbiotic Interaction. Symbiotic 2015. Lecture Notes in Computer Science(), vol 8820. Springer, Cham. https://doi.org/10.1007/978-3-319-13500-7_9
Download citation
DOI: https://doi.org/10.1007/978-3-319-13500-7_9
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-13499-4
Online ISBN: 978-3-319-13500-7
eBook Packages: Computer ScienceComputer Science (R0)