HCI Bibliography Home | HCI Conferences | About IUI | IUI Conf Proceedings | Detailed Records | RefWorks | EndNote | Hide Abstracts
IUI Tables of Contents: 939798990001020304050607080910111213-113-214-114-2 ⇐ MORE

Proceedings of the 2012 International Conference on Intelligent User Interfaces

Fullname:International Conference on Intelligent User Interfaces
Editors:Carlos Duarte; Luís Carriço; Joaquim Jorge; Sharon Oviatt; Daniel Gonçalves
Location:Lisbon, Portugal
Dates:2012-Feb-14 to 2012-Feb-17
Standard No:ISBN: 1-4503-1048-6, 978-1-4503-1048-2; ACM DL: Table of Contents hcibib: IUI12
Links:Conference Home Page
Summary:It is our great pleasure to welcome you to the 2012 ACM International Conference on Intelligent User Interfaces -- IUI'12 and to Lisboa, Portugal. Starting in 1993, IUI is now on its 17th edition and has established itself has the premier venue for reporting outstanding research and development on intelligent user interface. The IUI series continues to be the principal forum for the meeting of the Human-Computer Interaction (HCI) and the Artificial Intelligence (AI) communities. This gives rise to unique, creative, interdisciplinary contributions. This year's meeting showcases those topics, with works covering innovations in mobile interfaces, collaboration technology, affective interfaces, multimodal interfaces, pen interfaces, haptic and gesture interfaces, multi-touch interfaces, speech interfaces, gaze-based interfaces, tabletop interfaces, health applications, entertainment applications, adaptive narratives and theater, sports and in-vehicle applications, geographic applications, social media interfaces, educational interfaces, sketch recognition, human-robot interfaces, personalization and assistive technologies, end-user programming, ubiquitous and smart environments, locationaware interfaces, recommender interfaces, persuasive interfaces, web-based interfaces and agentbased interfaces.
    Building on previous years' success, the IUI 2012 call for papers attracted 134 full paper submissions and 78 short paper submissions. To ensure the highest possible quality we assembled a team of 48 senior program committee members and 367 reviewers. This guaranteed that every submission had at least three reviews plus one meta-review. This was followed by a rebuttal phase, keeping the procedure started two years ago. We believe that this process has guaranteed the highest quality possible for this year technical program and we would like to thank all the senior program committee members and reviewers for their hard work in making sure the most relevant works were selected. Through this very thorough review process we were able to accept 18 full papers, 15 short papers and 16 poster presentations, meaning an acceptance rate of 13% for full papers, 16% for oral presentations and 23% for combined oral and poster presentations.
    In addition to the full and short papers presentations and poster presentations, IUI will also feature a Demonstrations session for which 18 submissions were accepted from a total of 26 submissions.
    This year, the IUI programme includes three notable invited speakers sharing their innovative work and experiences with the conference participants: Alex 'Sandy' Pentland from the MIT Human Dynamic Lab, Christopher Bishop from Microsoft Research and Takeo Igarashi from the University of Tokyo.
    In order to foster the growth of the community, IUI 2012 will feature for the first time in the IUI series a Doctoral Consortium. Its success can be measured by the 12 high quality submissions received. Doctoral Consortium students will also be able to exhibit their work to the main conference audience in the poster session, thus promoting their insertion into the IUI community. We would also like to thank the conference sponsors and supporters that made it possible to provide financial support for Doctoral Consortium attendees and other students participating in the conference.
    Nine full-day workshops will take place at IUI 2012 covering diverse trending topics in the IUI area: Activity Context Representation: Techniques and Languages; Developing Intelligent User Interfaces for e-Accessibility and e-Inclusion; Context-awareness in Retrieval and Recommendation (CaRR 2012); 3rd Workshop on Semantic Models for Adaptive Interactive Systems (SEMAIS); Scent and Scensibility; User Modeling from Social Media; 2nd Workshop on Interacting with Smart Objects; 2nd Workshop on Location Awareness for Mixed and Dual Reality (LaMDa'12); 1st International Workshop on Ubiquitous Personalization (UP'2012).
    IUI 2012 continues and extends the cooperation with the ACM Transactions on Interactive Intelligent Systems (TiiS). Besides the streamlined processing of journal submissions based on IUI 2012 papers, IUI 2012 will feature a special session where authors of recently accepted articles for the ACM TiiS will spotlight key contributions of their work. It is our conviction that it will benefit greatly every member of the IUI community. We would like to thank Anthony Jameson and John Riedl, the Editors in Chief of TiiS, for their cooperation.
  1. Mobile interfaces & novel interaction
  2. Learning with technology
  3. Gestures and large displays
  4. Designing narratives & theater
  5. Personalization, search & usability
  6. Social computing & media
  7. Smart infrastructure
  8. Poster presentation
  9. Demonstration session
  10. Doctoral consortium
  11. Workshops

Mobile interfaces & novel interaction

Probabilistic pointing target prediction via inverse optimal control BIBAFull-Text 1-10
  Brian Ziebart; Anind Dey; J. Andrew Bagnell
Numerous interaction techniques have been developed that make "virtual" pointing at targets in graphical user interfaces easier than analogous physical pointing tasks by invoking target-based interface modifications. These pointing facilitation techniques crucially depend on methods for estimating the relevance of potential targets. Unfortunately, many of the simple methods employed to date are inaccurate in common settings with many selectable targets in close proximity. In this paper, we bring recent advances in statistical machine learning to bear on this underlying target relevance estimation problem. By framing past target-driven pointing trajectories as approximate solutions to well-studied control problems, we learn the probabilistic dynamics of pointing trajectories that enable more accurate predictions of intended targets.
Towards online adaptation and personalization of key-target resizing for mobile devices BIBAFull-Text 11-20
  Tyler Baldwin; Joyce Chai
Software (soft) keyboards are becoming increasingly popular on mobile devices. To attempt to improve soft keyboard input accuracy, key-target resizing algorithms that dynamically change the size of each key's target area have been developed. Although methods that employ personalized touch models have been shown to outperform general models, previous work has relied upon laboratory-based offline calibration to collect the data necessary to build these models. Such approaches are unrealistic and interuptive, and it is unlikely that offline calibration can be applied in a realistic usage setting, as hundreds or thousands of touch points are necessary to build the models. To combat this problem, this paper explores the possibility of online adaptation of key-target resizing algorithms. In particular, we propose and examine three online data collection methods that can be used to build and dynamically update personalized key-target resizing models. Our results suggest that a data collection methodology that makes inference based on vocabulary and error correction behavior is able to perform on par with gold standard personalized models, while reducing relative error rate by 10.4% over general models. This approach is simple, computationally inexpensive, and calculable via information that the system already has access to. Additionally, we show that these models can be built quickly, requiring less than one week's worth of text input by an average mobile device user.
Multimodal interaction patterns in mobile local search BIBAFull-Text 21-24
  Patrick Ehlen; Michael Johnston
Speak4it is a mobile search application that leverages multimodal input and integration to allow users to search for and act on local business information. We present an initial empirical analysis of user interaction with a multimodal local search application deployed in the field with real users. Specifically, we focus on queries involving multimodal commands, and analyze multimodal interaction behaviors seen in a deployed multimodal system.
In-vehicle driver recognition based on hand ECG signals BIBAFull-Text 25-28
  Hugo Silva; André Lourenço; Ana Fred
We present a system for in-vehicle driver recognition based on biometric information extracted from electrocardiographic (ECG) signals collected at the hands. We recur to non-intrusive techniques, that are easy to integrate into components with which the driver naturally interacts with, such as the steering wheel. This system is applicable to the automatic customization of vehicle settings according to the perceived driver, being also prone to expand the security features of the vehicle through the detection of hands-off steering wheel events in a continuous or near-continuous manner. We have performed randomized tests for performance evaluation of the system, in a subject identification scenario, using closed sets of up to 5 subjects, showing promising results for the intended application.
Performance comparisons of phrase sets and presentation styles for text entry evaluations BIBAFull-Text 29-32
  Per Ola Kristensson; Keith Vertanen
We empirically compare five different publicly-available phrase sets in two large-scale (N = 225 and N = 150) crowdsourced text entry experiments. We also investigate the impact of asking participants to memorize phrases before writing them versus allowing participants to see the phrase during text entry. We find that asking participants to memorize phrases increases entry rates at the cost of slightly increased error rates. This holds for both a familiar and for an unfamiliar text entry method. We find statistically significant differences between some of the phrase sets in terms of both entry and error rates. Based on our data, we arrive at a set of recommendations for choosing suitable phrase sets for text entry evaluations.
Towards providing just-in-time vocabulary support for assistive and augmentative communication BIBAFull-Text 33-36
  Carrie Demmans Epp; Justin Djordjevic; Shimu Wu; Karyn Moffatt; Ronald M. Baecker
Many people cannot communicate effectively with those around them. The causes vary but several tools and strategies can support their communication. These tools, which collectively fall under the banner of Assistive and Augmentative Communication (AAC), are rarely adaptive. Of those that are, few provide context-based or just-in-time vocabulary support to users even though the proliferation of smartphones makes this possible. To meet this need, we developed four algorithms to retrieve relevant vocabulary from Internet-based corpora. We used discourse completion tasks to evaluate each algorithm's ability to identify appropriate vocabulary across a set of specific contexts. The results indicate that our approach identifies appropriate context-specific words that complement general AAC vocabularies: when combined with a typical base vocabulary, the algorithms outperformed the support provided by the base vocabulary alone. They did this by adding small targeted vocabularies.
Mobile texting: can post-ASR correction solve the issues? an experimental study on gain vs. costs BIBAFull-Text 37-40
  Michael Feld; Saeedeh Momtazi; Farina Freigang; Dietrich Klakow; Christian Müller
The next big step in embedded, mobile speech recognition will be to allow completely free input as it is needed for messaging like SMS or email. However, unconstrained dictation remains error-prone, especially when the environment is noisy. In this paper, we compare different methods for improving a given free-text dictation system used to enter textbased messages in embedded mobile scenarios, where distraction, interaction cost, and hardware limitations enforce strict constraints over traditional scenarios. We present a corpus-based evaluation, measuring the trade-off between improvement of the word error rate versus the interaction steps that are required under various parameters. Results show that by post-processing the output of a "black box" speech recognizer (e.g. a web-based speech recognition service), a reduction of word error rate by 55% (10.3% abs.) can be obtained. For further error reduction, however, a richer representation of the original hypotheses (e.g. lattice) is necessary.

Learning with technology

Style by demonstration: teaching interactive movement style to robots BIBAFull-Text 41-50
  James Young; Kentaro Ishii; Takeo Igarashi; Ehud Sharlin
The style in which a robot moves, expressed through its gait or locomotion, can convey effective messages to people. For example, a robot could move aggressively in reaction to a person's actions, or alternatively react using a set of careful, submissive movements. Designing, implementing and programming robotic interfaces that react to users' actions with properly styled movements can be a difficult, daunting, and time consuming technical task. On the other hand, most people can easily perform such stylistic tasks and movements, for example, through acting them out.
   Following this observation, we propose to enable people to use their existing teaching skills to directly demonstrate to robots, via in-situ acting, a desired style of interaction. In this paper we present an initial style-by-demonstration (SBD) proof-of-concept of our approach, allowing people to teach a robot specific, interactive locomotion styles by providing a demonstration. We present a broomstick-robot interface for directly demonstrating locomotion style to a collocated robot, and a design critique evaluation by experienced programmers that compares our SBD approach to traditional programming methods.
PhysicsBook: a sketch-based interface for animating physics diagrams BIBAFull-Text 51-60
  Salman Cheema; Joseph LaViola
We present PhysicsBook, a prototype system that enables users to solve physics problems using a sketch-based interface and then animates any diagram used in solving the problem to show that the solution is correct. PhysicsBook recognizes the diagrams in the solution and infers relationships among diagram components through the recognition of mathematics and annotations such as arrows and dotted lines. For animation, PhysicsBook uses a customized physics engine that provides entry points for hand-written mathematics and diagrams. We discuss the design of PhysicsBook, including details of algorithms for sketch recognition, inference of user intent and creation of animations based on the mathematics written by a user. Specifically, we describe how the physics engine uses domain knowledge to perform data transformations in instances where it cannot use a given equation directly. This enables PhysicsBook to deal with domains of problems that are not directly related to classical mechanics. We provide examples of scenarios of how PhysicsBook could be used as part of an intelligent tutoring system and discuss the strengths and weaknesses of our current prototype. Lastly, we present the findings of a preliminary usability study with five participants.
A learning support tool with clinical cases based on concept maps and medical entity recognition BIBAFull-Text 61-70
  Manuel de la Villa; Fernando Aparicio; Manuel J. Maña; Manuel de Buenaga
The search for truthful health information through Internet is an increasingly complex process due to the growing amount of resources. Access to information can be difficult to control even in environments where the goal pursued is well-defined, as in the case of learning activities with medical students. In this paper, we present a computer tool devised to ease the process of understanding medical concepts from information in clinical case histories. To this end, it automatically constructs concept maps and presents reliable information from different ontologies and knowledge bases. The two main components of the system are an Intelligent Information Access interface and a Concept Map Graph that retrieves medical concepts from a text input, and provides rich information and semantically related concepts. The paper includes a user evaluation of the first component and a systematic assessment for the second component. Results show that our proposal can be efficient and useful for students in a medical learning environment.
PedConnect: an intelligent assistant for teacher social networking BIBAFull-Text 71-74
  Jihie Kim; Yu-Han Chang; Sen Cai; Siddharth Jain
Social networking has gained immense traction in many areas, including teaching and learning. Networking sites for teachers aim to facilitate teacher communication and information sharing, but fall short of their potential. In order to support more effective use of online resources and better communication among teachers, we develop a suite of new user modeling and recommendation capabilities within a middle school teacher networking site. We foster collaboration among novice and experienced teachers when they share similar interests, enabling new mentoring relationships, and promote the use of relevant educational resources. We illustrate our approach with an implemented system called PedConnect that analyzes user activities and presents intelligent suggestions for collaboration and resource use.

Gestures and large displays

An adaptive solution for intra-operative gesture-based human-machine interaction BIBAFull-Text 75-84
  Ali Bigdelou; Loren Schwarz; Nassir Navab
Computerized medical systems play a vital role in the operating room, however, sterility requirements and interventional workflow often make interaction with these devices challenging for surgeons. Typical solutions, such as delegating physical control of keyboard and mouse to assistants, add an undesirable level of indirection. We present a touchless, gesture-based interaction framework for the operating room that lets surgeons define a personalized set of gestures for controlling arbitrary medical computerized systems. Instead of using cameras for capturing gestures, we rely on a few wireless inertial sensors, placed on the arms of the surgeon, eliminating the dependence on illumination and line-of-sight. A discriminative gesture recognition approach based on kernel regression allows us to simultaneously classify performed gestures and to track the relative spatial pose within each gesture, giving surgeons fine-grained control of continuous parameters. An extensible software architecture enables a dynamic association of learned gestures to arbitrary intraoperative computerized systems. Our experiments illustrate the performance of our approach and encourage its practical applicability.
Unistroke gestures on multi-touch interaction: supporting flexible touches with key stroke extraction BIBAFull-Text 85-88
  Yingying Jiang; Feng Tian; Xiaolong Zhang; Wei Liu; Guozhong Dai; Hongan Wang
Gesture inputs on multi-touch tabletops usually involve multiple fingers (more than two) and casual touchdowns or liftoffs of fingers. This flexibility of touch gestures allows more natural user interaction, but also poses new challenges for accurate recognition of multi-touch gestures. To address these challenges, we propose a new approach to recognize flexible multi-touch stroke gestures on tabletops. Based on a user study on multi-touch unistroke gestures, we develop a gesture recognition method by extracting key strokes embedded in flexible multi-touch input. Our evaluation study result shows that this method can greatly improve the recognition accuracy of flexible multi-touch unistroke gestures on tabletops.
Continuous recognition of one-handed and two-handed gestures using 3D full-body motion tracking sensors BIBAFull-Text 89-92
  Per Ola Kristensson; Thomas Nicholson; Aaron Quigley
In this paper we present a new bimanual markerless gesture interface for 3D full-body motion tracking sensors, such as the Kinect. Our interface uses a probabilistic algorithm to incrementally predict users' intended one-handed and two-handed gestures while they are still being articulated. It supports scale and translation invariant recognition of arbitrarily defined gesture templates in real-time. The interface supports two ways of gesturing commands in thin air to displays at a distance. First, users can use one-handed and two-handed gestures to directly issue commands. Second, users can use their non-dominant hand to modulate single-hand gestures. Our evaluation shows that the system recognizes one-handed and two-handed gestures with an accuracy of 92.7%-96.2%.
Studying user-defined iPad gestures for interaction in multi-display environment BIBAFull-Text 93-96
  Ekaterina Kurdyukova; Matthias Redlin; Elisabeth André
The paper investigates the iPad gestures that users naturally perform for data transfer. We examine the transfer between two iPads, iPad and a tabletop, and iPad and a public display. Three gesture modalities are investigated: multi-touch gestures, performed using iPad display, spatial gestures, performed by manipulating iPad in 3D space, and direct contact gestures, involving the physical contact of iPad and other device. We report on user choices of the modalities and gesture types, and derive critical points for the design of iPad gestures.
Using intelligent natural user interfaces to support sales conversations BIBAFull-Text 97-100
  Sven Gehring; Markus Löchtefeld; Florian Daiber; Matthias Böhmer; Antonio Krüger
During sales conversations, gestures and mimics are of high importance to communicate information about a product. One prominent example for such sales gestures is the meat and cheese counter, which is one of the remaining spots in supermarkets where sales persons interact with customers. Interactions at such counters in supermarkets normally follow a simple protocol. The customer points at an item of choice. The employee takes out the item and, in most of the cases the product needs to be cut to fit the amount the customer wants to buy. Often it is ambiguous about what specific product the customer and the employees are talking about. Up to now, there are just a few efforts in HCI research to enrich communication at the point of sale. In this paper we report and analyze one scenario in which an intelligent natural user interface can support communication between customer and employee in a sales conversation. Furthermore, we report on our prototype that is able to track pointing gestures by using a depth camera and to display information about items pointed at.
A glove for tapping and discrete 1D/2D input BIBAFull-Text 101-104
  Sam Miller; Andy Smith; Sina Bahram; Robert St. Amant
This paper describes a glove with which users enter input by tapping fingertips with the thumb or by rubbing the thumb over the palmar surfaces of the middle and index fingers. The glove has been informally tested as the controller for two semi-autonomous robots in a 3D simulation environment. A preliminary evaluation of the glove's performance is presented.
Using overlays to support collaborative interaction with display walls BIBAFull-Text 105-108
  Arvind Satyanarayan; Nadir Weibel; James Hollan
Large-scale display walls, and the high-resolution visualizations they support, promise to become ubiquitous. Natural interaction with them, especially in collaborative environments, is increasingly important and yet remains an on-going challenge. Part of the problem is a resolution mismatch between low-resolution input devices and high-resolution display walls. In addition, enabling concurrent use by multiple users is difficult -- for example, how would this large workspace be managed for multiple users and what novel collaborative interactions could occur? In this paper, we present an overlay interface element superimposed on wall-display applications to help constrain interaction, focus attention on subsections of a display wall, and facilitate a collaborative multi-user workflow.

Designing narratives & theater

Virtual stage linked with a physical miniature stage to support multiple users in planning theatrical productions BIBAFull-Text 109-118
  Yosuke Horiuchi; Tomoo Inoue; Ken-ichi Okada
Theater is a collaborative art form that involves production team members with different specialties. Because theater involves various technical elements, such as stage design and lighting, the production team must work in cooperation among various departments to design a theatrical production. When planning a theatrical production, it is difficult to visualize the stage as a whole and to incorporate the ideas of production team members from various departments. In this paper, we propose a system for reproducing the theatrical stage by means of a virtual stage linked to a physical miniature stage. The miniature stage is presented on a tabletop interface, and the virtual stage is created by computer graphics to reflect the actions on the miniature stage in real time. By actually presenting theatrical production ideas in two spaces, users can more easily collaborate and gain a comprehensive view of the stage.
Exploring passive user interaction for adaptive narratives BIBAFull-Text 119-128
  Stephen Gilroy; Julie Porteous; Fred Charles; Marc Cavazza
Previous Interactive Storytelling systems have been designed to allow active user intervention in an unfolding story, using established multi-modal interactive techniques to influence narrative development. In this paper we instead explore the use of a form of passive interaction where users' affective responses, measured by physiological proxies, drive a process of narrative adaptation. We introduce a system that implements a passive interaction loop as part of narrative generation, monitoring users' physiological responses to an on-going narrative visualization and using these to adapt the subsequent development of character relationships, narrative focus and pacing. Idiomatic cinematographic techniques applied to the visualization utilize existing theories of establishing characteristic emotional tone and viewer expectations to foster additional user response. Experimental results support the applicability of filmic emotional theories in a non-film visual realization, demonstrating significant appropriate user physiological response to narrative events and "emotional cues". The subsequent narrative adaptation provides a variation of viewing experience with no loss of narrative comprehension.

Personalization, search & usability

On slide-based contextual cues for presentation reuse BIBAFull-Text 129-138
  Moushumi Sharmin; Lawrence Bergman; Jie Lu; Ravi Konuru
Reuse of existing presentation materials is prevalent among knowledge workers. However, finding the most appropriate material for reuse is challenging. Existing information management and search tools provide inadequate support for reuse due to their dependence on users' ability to effectively categorize, recall, and recognize existing materials. Based on our findings from an online survey and contextual interviews, we designed and implemented a slide-based contextual recommender, ConReP, for supporting reuse of presentation materials. ConReP utilizes a user-selected slide as a search-key, recommends materials based on similarity to the selected slide, and provides a local-context-based visual representation of the recommendations. Users input provides new insight into presentation reuse and reveals that slide-based search is more effective than keyword-based search, local-context-based visual representation helps in better recall and recognition, and shows the promise of this general approach of exploiting individual slides and local-context for better presentation reuse.
Collecting relevance feedback on titles and photographs in weblog posts BIBAFull-Text 139-148
  Amy Campbell; Christopher Wienberg; Andrew Gordon
We investigate new interfaces that allow users to specify topics of interest in streams of weblog stories by providing relevance feedback to a search algorithm. Noting that weblog stories often contain photographs taken by the blogger during the course of the narrated events, we investigate whether these photographs can serve as a proxy for the whole post when users are making judgments as to the post's relevance. We developed a new story annotation interface for collecting relevance feedback with three variations: users are presented either with the full post as it appears in a weblog, an embedded photograph, or only the title of the post. We describe a user evaluation that compares annotation time, quality, and subjective user experience across each of these three conditions. The results show that relevance judgments based on embedded photographs or titles are far less accurate than when reading the whole weblog post, but the time required to acquire an accurate model of the user's topic interest is greatly reduced.
Effective browsing and serendipitous discovery with an experience-infused browser BIBAFull-Text 149-158
  Sudheendra Hangal; Abhinay Nagpal; Monica Lam
In the digital age, users can have perfect recall of their online experiences. In this paper, we explore how this recall can be leveraged during web browsing.
   We have built a system called the Experience-Infused Browser that indexes a user's digital history such as email and chat archives. As the user browses the web, it observes the contents of pages viewed, and appropriately highlights named entities on the page that the user has encountered in the past. This browser has two benefits. First, it highlights terms on the page that occur frequently in the user's communications, effectively personalizing the page for the user. Second, the system can remind the user of names that he has encountered in the past but may not remember.
   We evaluated how users reacted to the browser during organic web browsing. Our users have reported that it was useful on crowded web pages to surface content that they otherwise may have missed, and in recalling serendipitous connections to people that they had forgotten. Most of our users said they would use the browser beyond the experimental study, indicating that they derived clear benefit from it.
Towards an intelligent reviewer's assistant: recommending topics to help users to write better product reviews BIBAFull-Text 159-168
  Ruihai Dong; Kevin McCarthy; Michael O'Mahony; Markus Schaal; Barry Smyth
User opinions and reviews are an important part of the modern web and all major e-commerce sites typically provide their users with the ability to provide and access customer reviews across their product catalog. Indeed this has become a vital part of the service provided by sites like Amazon and TripAdvisor, so much so that many of us will routinely check appropriate product reviews before making a purchase decision, regardless of whether we intend to purchase online or not. The importance of reviews has highlighted the need to help users to produce better reviews and in this paper we describe the development and evaluation of a Reviewer's Assistant for this purpose. We describe a browser plugin that is designed to work with major sites like Amazon and to provide users with suggestions as they write their reviews. These suggestions take the form of topics (e.g. product features) that a reviewer may wish to write about and the suggestions automatically adapt as the user writes their review. We describe and evaluate a number of different algorithms to identify useful topics to recommend to the user and go on to describe the results of a preliminary live-user trial.
Are explanations always important?: a study of deployed, low-cost intelligent interactive systems BIBAFull-Text 169-178
  Andrea Bunt; Matthew Lount; Catherine Lauzon
Intelligent interactive systems (IIS) have great potential to improve users' experience with technology by tailoring their behaviour and appearance to users' individual needs; however, these systems, with their complex algorithms and dynamic behaviour, can also suffer from a lack of comprehensibility and transparency. We present the results of two studies examining the comprehensibility of, and desire for explanations with deployed, low-cost IIS. The first study, a set of interviews with 21 participants, reveals that i) comprehensibility is not always dependent on explanations, and ii) the perceived cost of viewing explanations tends to outweigh the anticipated benefits. Our second study, a two-week diary study with 14 participants, confirms these findings in the context of daily use, with participants indicating a desire for an explanation in only 7% of diary entries. We discuss the implications of our findings for the design of explanation facilities.

Social computing & media

Modeling topic specific credibility on twitter BIBAFull-Text 179-188
  Byungkyu Kang; John O'Donovan; Tobias Höllerer
This paper presents and evaluates three computational models for recommending credible topic-specific information in Twitter. The first model focuses on credibility at the user level, harnessing various dynamics of information flow in the underlying social graph to compute a rating. The second model applies a content-based strategy to compute a finer-grained credibility score for individual tweets. Lastly, we discuss a third model which combines facets from both models in a hybrid method, using both averaging and filtering hybrid strategies. To evaluate our novel credibility models, we perform an evaluation on 7 topic specific data sets mined from the Twitter streaming API, with specific focus on a data set of 37K users who tweeted about the topic "Libya". Results show that the social model outperfoms hybrid and content-based prediction models in terms of predictive accuracy over a set of manually collected credibility ratings on the "Libya" dataset.
Summarizing sporting events using twitter BIBAFull-Text 189-198
  Jeffrey Nichols; Jalal Mahmud; Clemens Drews
The status updates posted to social networks, such as Twitter and Facebook, contain a myriad of information about what people are doing and watching. During events, such as sports games, many updates are sent describing and expressing opinions about the event. In this paper, we describe an algorithm that generates a journalistic summary of an event using only status updates from Twitter as a source. Temporal cues, such as spikes in the volume of status updates, are used to identify the important moments within an event, and a sentence ranking method is used to extract relevant sentences from the corpus of status updates describing each important moment within an event. We evaluate our algorithm compared to human-generated summaries and the previous best summarization algorithm, and find that the results of our method are superior to the previous algorithm and approach the readability and grammaticality of the human-generated summaries.
EPIC: a multi-tiered approach to enterprise email prioritization BIBAFull-Text 199-202
  Jie Lu; Zhen Wen; Shimei Pan; Jennifer Lai
We present Enterprise Priority Inbox Classifier (EPIC), an automatic personalized email prioritization system based on a topic-based user model built from the user's email data and relevant enterprise information. The user model encodes the user's topics of interest and email processing behaviors (e.g. read/reply/file) at the granularity of pair-wise interactions between the user and each of his/her email contacts. Given a new message, the user model is used in combination with the message metadata and content to determine the values of a set of contextual features. Contextual features include people-centric features representing information about the user's interaction history and relationship with the email sender, as well as message-centric features focusing on the properties of the message itself. Based on these feature values, EPIC uses a dynamic strategy to combine a global priority classifier with a user-specific classifier for determining the message's priority. An evaluation of EPIC based on 2,064 annotated email messages from 11 users, using 10-fold cross-validation, showed that the system achieves an average accuracy of 81.3%. The user-specific classifier contributed an improvement of 11.5%. Lastly we report on findings regarding the relative value of different contextual features for email prioritization.
Finding someone in my social directory whom i do not fully remember or barely know BIBAFull-Text 203-206
  Michelle Zhou; Wei Zhang; Barton Smith; Erika Varga; Martin Farias; Hernan Badenes
REACH is an intelligent, people-finding system that helps users to find someone in their social directory, especially those whom they do not fully remember or barely know. It analyzes a user's communication and social networking data to automatically extract all the contacts and derive multiple facets to characterize each contact in relation to the user. It then employs a personalized, faceted search to retrieve and present a ranked list of matched contacts based on their properties. A preliminary evaluation shows the effectiveness of our approach.

Smart infrastructure

A visual programming language for designing interactions embedded in web-based geographic applications BIBAFull-Text 207-216
  The Nhan Luong; Patrick Etcheverry; Christophe Marquesuzaà; Thierry Nodenot
Visual programming languages (VPLs) provide notations for representing both the intermediate and the final results of a knowledge engineering process. Whereas some VPLs particularly focus on control flow and/or data flow of a software, very few VPLs stress on the interactive dimension of application (dialogue flow). This paper focuses on a VPL allowing designers to specify interactions between a user and a system, in the field of Web-based geographic applications. We first present the underlying interaction model that the VPL is based on, and then the detailed characteristics of the VPL. We show how this VPL has been integrated in a graphical design framework allowing designers to immediately assess their specification. Then we illustrate the way to use the framework from the design step to the final code generation step. Last, we detail an experimentation aiming at evaluating the strengths and the weaknesses of our VPL.
Automatic reverse engineering of interactive dynamic web applications to support adaptation across platforms BIBAFull-Text 217-226
  Federico Bellucci; Giuseppe Ghiani; Fabio Paternò; Claudio Porta
The effort and time required to develop user interface models has been one of the main limitations to the adoption of model-based approaches, which enable intelligent processing of user interface descriptions. In this paper, we present a tool to perform reverse engineering of interactive dynamic Web applications into a model-based framework able to describe them at various abstraction levels. We indicate how information in HTML, HTML 5, CSS, Ajax and JavaScript is transformed into such logical framework, which facilitates adaptation to other types of interactive devices. We also discuss how this reverse engineering tool has been exploited in an environment for run-time adaptation or migration of interactive Web applications to various devices in ubiquitous use cases.
Towards automatic functional test execution BIBAFull-Text 227-236
  Pablo Pedemonte; Jalal Mahmud; Tessa Lau
As applications are developed, functional tests ensure they continue to function as expected. Nowadays, functional testing is mostly done manually, with human testers verifying a system's functionality themselves, following hand-written instructions. While there exist tools supporting functional test automation, in practice they are hard to use, require programming skills, and do not provide good support for test maintenance. In this paper, we take an alternative approach: we semi-automatically convert hand-written instructions into automated tests. Our approach consists of two stages: first, employing machine learning and natural language processing to compute an intermediate representation from test steps; and second, interactively disambiguating that representation to create a fully automated test. These two stages comprise a complete system for converting hand-written functional tests into automated tests. We also present a quantitative study analyzing the effectiveness of our approach. Our results show that 70% of manual test steps can be automatically converted to automated test steps with no user intervention.
Lin-spiration: using a mixture of spiral and linear visualization layouts to explore time series BIBAFull-Text 237-240
  Eduardo Graells; Alejandro Jaimes
Time series data is pervasive in many domains and interactive visualization of such data is useful for a wide range of tasks including analysis and prediction. In spite of the importance of visualizing time series data and the fact that time series data is often easily interpretable, traditional approaches are either very simple and limited, or are aimed at domain experts. In this paper, we propose a novel interactive visualization paradigm for exploring and comparing multiple sets of time series data. In particular, we propose a focus+context approach, where a "focus" segment of a time series is zoomed into and visualized using a linear layout at one scale, while the remaining segments of the time series (i.e., the context) are visualized using spiral data layouts. Our paradigm allows the user to dynamically select and compare different sections of each time series independently, facilitating the exploration of time series data in a fun and engaging way.

Poster presentation

Intelligent pairing assistant for air operation centers BIBAFull-Text 241-244
  Jeremy Ludwig; Eric Geiselman
Within an Air Operations Center (AOC), planners make crucial decisions to create the air plan for any given day. They are expected to complete the plan in part by pairing targeting or collection tasks with the available platforms. Any assistance these planners can acquire to help create the plan in a timely manner would make the entire process more efficient and effective. This paper describes the Intelligent Pairing Assistant (IPA) prototype, which would provide pairing recommendations at specific decision points in the planning process. IPA is designed as a plug-in for software systems already in use within AOCs. The primary contribution described in this paper is the application of existing research in intelligent user interfaces to a novel domain.
CAVIAR: a vibrotactile device for accessible reaching BIBAFull-Text 245-248
  Sina Bahram; Arpan Chakraborty; Robert St. Amant
CAVIAR is designed to aid people with vision impairment in locating, identifying, and acquiring objects in their peripersonal space. A mobile phone, worn on the chest, captures video in front of the user; the computer vision component locates the user's hand and objects in the video stream. The auditory component informs the user about the presence of objects. On user confirmation, the reaching component sends signals to vibrotactile actuators on the user's wristband, guiding the hand to a specific object. This paper describes an end-to-end prototype of CAVIAR and its formative evaluation.
Clustering web pages to facilitate revisitation on mobile devices BIBAFull-Text 249-252
  Jie Liu; Chun Yu; Wenchang Xu; Yuanchun Shi
Due to small screens, inaccuracy of input and other limitations of mobile devices, revisitation of Web pages in mobile browsers takes more time than that in desktop browsers. In this paper, we propose a novel approach to facilitate revisitation. We designed AutoWeb, a system that clusters opened Web pages into different topics based on their contents. Users can quickly find a desired opened Web page by narrowing down the searching scope to a group of Web pages that share the same topic. Clustering accuracy is evaluated to be 92.4% and computing resource consumption was proved to be acceptable. A user study was conducted to explore user experience and how much AutoWeb facilitates revisitation. Results showed that AutoWeb could save up a significant time for revisitation and participants rated the system highly.
Out of the bubble: serendipitous even recommendations at an urban music festival BIBAFull-Text 253-256
  Andreas Forsblom; Petteri Nurmi; Pirkka Åman; Lassi Liikkanen
Advances in positioning technologies have resulted in a surge of location-based recommendation systems for mobile devices. A central challenge in these systems is to avoid the so-called filter bubble effect, i.e., that people are not only exposed to information that is in line with their personal ecosystem, but that they can also discover novel and otherwise interesting content. We present results from a field study of a mobile recommendation system that has been aimed to support serendipitous discovery of events at an urban culture festival. Results from the study indicate that suitably designed recommendations together with access to relevant external information sources can lead to serendipitous discovery of new content, such as new artists, bands or individual songs. Our results also indicate that proximity has little effect on the effectiveness of serendipitous recommendations.
Image registration for text-gaze alignment BIBAFull-Text 257-260
  Pascual Martinez-Gomez; Chen Chen; Tadayoshi Hara; Yoshinobu Kano; Akiko Aizawa
Applications using eye-tracking devices need a higher accuracy in recognition when the task reaches a certain complexity. Thus, more sophisticated methods to correct eye-tracking measurement errors are necessary to lower the penetration barrier of eye-trackers in unconstrained tasks. We propose to take advantage of the content or the structure of textual information displayed on the screen to build informed error-correction algorithms that generalize well. The idea is to use feature-based image registration techniques to perform a linear transformation of gaze coordinates to find a good alignment with text printed on the screen. In order to estimate the parameters of the linear transformation, three optimization strategies are proposed to avoid the problem of local minima, namely Monte Carlo, multi-resolution and multi-blur optimization. Experimental results show that a more precise alignment of gaze data with words on the screen can be achieved by using these methods, allowing a more reliable use of eye-trackers in complex and unconstrained tasks.
Music listening history explorer: an alternative approach for browsing music listening history habits BIBAFull-Text 261-264
  Ricardo Dias; Manuel J. Fonseca; Daniel Gonçalves
Nowadays, people spend time using services to track their music listening history. Although these services provide statistics and small graphics/charts, they are mainly used to record and to allow direct access to the information, not providing any visualization and exploration functionality. In this paper we describe a new approach for browsing and visualizing music listening histories, which combines a timeline-based visualization, with a set of synchronized-views and an interactive filtering mechanism to provide a flexible and easy to use solution. This was complemented with brushing and highlighting techniques that allow users to observe trends on artists, albums and tracks listening. Experimental evaluation with users revealed that they were able to complete all the proposed tasks with a low error rate, and that they found the solution easy to use. Moreover, users liked our approach for browsing and exploring listening histories, emphasizing its flexibility and effectiveness, and founding the full experience engaging and rewarding.
LogicPad: a pen-based application for visualization and verification of boolean algebra BIBAFull-Text 265-268
  Bo Kang; Joseph LaViola
We present LogicPad, a pen-based application for boolean algebra visualization that lets users manipulate boolean function representations through handwritten symbol and gesture recognition coupled with a drag-and-drop interface. We discuss LogicPad's user interface and the general algorithm used for verifying the equivalence of three different boolean function representations: boolean expressions, truth tables, and logic gate diagrams. We also conducted a short, informal user study evaluating LogicPad's user interface, visualization techniques, and overall performance. Results show that visualizations were generally well-liked and verification results matched user expectations.
Interactive uncertainty analysis BIBAFull-Text 269-272
  Pedro Szekely; Yu-Han Chang; Rajiv Maheswaran; Yan Wang; Huihui Cheng; Karan Singh
Humans have difficulty evaluating the effects of uncertainty on schedules. People often mitigate the effects of uncertainty by adding slack based on experience and non-stochastic analyses such as the critical path method (CPM). This is costly as it leads to longer than necessary schedules, and can be ineffective without a clear understanding of where slack is needed. COMPASS is an interactive real-time tool that analyzes schedule uncertainty for a stochastic task network. An important feature is that it concurrently calculates stochastic critical paths and critical tasks. COMPASS visualizes this information on top of a traditional Gantt view, giving users insight into how delays caused by uncertain durations propagate down the schedule. Evaluations with 10 users show that users can use COMPASS to answer a variety of questions about the possible evolutions of a schedule (e.g., what is the likelihood that all activities will complete before a given date?)
The quest for validated personality trait stories BIBAFull-Text 273-276
  Matt Dennis; Judith Masthoff; Chris Mellish
This paper describes how a set of stories, each conveying a personality trait from the Five Factor Model at a high or low level, were developed using Amazon's Mechanical Turk. These stories will be used to develop interfaces that adapt to personality using a User as Wizard method. The paper shows how difficult it is to construct stories that convey a single personality trait. It also shows how such stories can be constructed for most cases, and how Mechanical Turk can aid to achieve this.
Audio cloud: creation and rendering BIBAFull-Text 277-280
  Jitendra Ajmera; Om D Deshmukh; Anupam Jain; Amit Anil Nanavati; Nitendra Rajput; Saurabh Srivastava
Word clouds are extensively used to present a summary of the prominent words in a document on the World Wide Web. Such clouds give the user an idea about the content of the document. In this paper we present a mechanism to create and render an audio cloud for audio content. Such audio clouds are expected to provide a similar summary of the audio documents. They have wide applicability in various domains, especially for low-literate users who currently do not use the Internet but interact with audio-based systems.
   Detecting words from an audio content is challenging, especially if the audio is in languages for which a speech recognition system does not exist. We present a language-independent mechanism to detect frequently occurring words within an audio document. We then present four ways to render these words that form an audio cloud. The four prototypes for rendering the audio cloud are based on varying the amplitude, the voice quality, echo and the repetition of audio words. An evaluation study conducted across 32 users suggests that literate and low-literate users easily understand the concept of audio cloud.
Pointing at responsive objects outdoors BIBAFull-Text 281-284
  YangLei Zhao; Arpan Chakraborty; Kyung Wha Hong; Shishir Kakaraddi; Robert St. Amant
In this paper we analyze pointing techniques for simple remote control of nearby and distant objects in an outdoor environment, using a mobile phone. In an experiment we determine the accuracy of pointing at targets from a few meters to a few hundred meters away, either by focusing the phone's camera on a target or holding the phone at waist level in the direction of the target. We describe a simulated network application in which users can activate and control one or more responsive objects using either interaction technique.
Towards recognizing "cool": can end users help computer vision recognize subjective attributes of objects in images? BIBAFull-Text 285-288
  William Curran; Travis Moore; Todd Kulesza; Weng-Keen Wong; Sinisa Todorovic; Simone Stumpf; Rachel White; Margaret Burnett
Recent computer vision approaches are aimed at richer image interpretations that extend the standard recognition of objects in images (e.g., cars) to also recognize object attributes (e.g., cylindrical, has-stripes, wet). However, the more idiosyncratic and abstract the notion of an object attribute (e.g., cool car), the more challenging the task of attribute recognition. This paper considers whether end users can help vision algorithms recognize highly idiosyncratic attributes, referred to here as subjective attributes. We empirically investigated how end users recognized three subjective attributes of carscool, cute, and classic. Our results suggest the feasibility of vision algorithms recognizing subjective attributes of objects, but an interactive approach beyond standard supervised learning from labeled training examples is needed.
Where do Facebook intelligent lists come from? BIBAFull-Text 289-292
  Fatoumata Camara; Gaëlle Calvary; Rachel Demumieux; Nadine Mandran
On September 19th 2011, Facebook introduced "Intelligent Lists" which are Friends Lists (FL) automatically created and pre-filled based on users' and their contacts' profiles information (education, work, city of living, kin, etc.). In early 2011, we conducted a study on contact management in Facebook in order to understand users' real needs. Outcomes from this study suggest several recommendations, some of which can be found today in the Facebook Intelligent Lists.
   This paper provides explanations on the recent evolution in Facebook contact management. The user study involved 148 participants. From their Facebook accounts, we retrieved 340 Friends Lists and 347 family ties. In the overall, the study has led to numerous interesting outcomes. In this paper, we focus on those related to Friends Lists and, particularly, on recommendations that have not yet been implemented in Facebook.
Machine listening: acoustic interface with ART BIBAFull-Text 293-296
  Benjamin Smith; Guy Garnett
Recent developments in machine listening present opportunities for innovative new paradigms for computer-human interaction. Voice recognition systems demonstrate a typical approach that conforms to event oriented control models. However, acoustic sound is continuous, and highly dimensional, presenting a rich medium for computer interaction. Unsupervised machine learning models present great potential for real-time machine listening and understanding of audio and sound data. We propose a method for harnessing unsupervised machine learning algorithms, Adaptive Resonance Theory specifically, in order to inform machine listening, build musical context information, and drive real-time interactive performance systems. We present the design and evaluation of this model leveraging the expertise of trained, improvising musicians.
1F: one accessory feature design for gesture recognizers BIBAFull-Text 297-300
  Radu-Daniel Vatavu
One Feature (1F) is a simple and intuitive pruning strategy that reduces considerably the amount of computations required by Nearest-Neighbor gesture classifiers while still preserving the high recognition rate. Performance results are reported for 1F by analyzing a large set of candidate features showing recognition rates of 99% with a peak reduction in computations of 70%. 1F is easy to implement, flexible with respect to the choice of the feature, and exploits the intuition of the designer by exposing clear inner workings.
Activity-aware recommendation for collaborative work in operating rooms BIBAFull-Text 301-304
  Afsaneh Doryab; Julian Togelius; Jakob Bardram
This paper presents a recommender system for teams of medical professionals working collaboratively in hospital operating rooms. The system recommends relevant virtual actions, such as retrieval of information resources and initiation of communication with professionals outside the operating rooms. Recommendations are based on the current state of the ongoing operation as recognised from sensor data using machine learning techniques. The selection and non-selection of virtual actions during operations are interpreted as implicit feedback and used to update the weight matrices that guide recommendations. A pilot user study involving medical professionals indicates that the adaptation mechanism is effective and that the system provides adequate recommendations.

Demonstration session

Universal earphones: earphones with automatic side and shared use detection BIBAFull-Text 305-306
  Kohei Matsumura; Daisuke Sakamoto; Masahiko Inami; Takeo Igarashi
We present universal earphones that use both a proximity sensor and a skin conductance sensor and we demonstrate several implicit interaction techniques they achieve by automatically detecting the context of use. The universal earphones have two main features. The first involves detecting the left and right sides of ears, which provides audio to either ear, and the second involves detecting the shared use of earphones and this provides mixed stereo sound to both earphones. These features not merely free users from having to check the left and right sides of earphones, but they enable them to enjoy sharing stereo audio with other people.
A demo of a facial UI design approach for digital artists BIBAFull-Text 307-308
  Pedro Bastos; Xenxo Alvarez Blanco; Verónica Orvalho
In the character animation industry, animators use facial UI's to animate a character's face. A facial UI provides widgets and handles that the animator interacts with to control the character's facial regions. This paper presents a facial UI design approach to control the animation of the six basic facial expressions of the anthropomorphic face. The design is based in square shaped widgets holding circular handles that allow the animator to produce the muscular activity relative to the basic facial expressions. We have implemented a prototype of the facial UI design in the Blender open-source animation software and did a preliminary pilot study with three animators. Two parameters were evaluated: the number of clicks and the time taken to animate the six basic facial expressions. The study reveals there was little variation in the values each animator marked for both parameters, despite the natural difference in their creative performance.
Simple, fast, and accurate clustering of data sequences BIBAFull-Text 309-310
  Luis Leiva; Enrique Vidal
Many devices generate large amounts of data that follow some sort of sequentiality, e.g., motion sensors, e-pens, or eye trackers, and therefore these data often need to be compressed for classification, storage, and/or retrieval purposes. This paper introduces a simple, accurate, and extremely fast technique inspired by the well-known K-means algorithm to properly cluster sequential data. We illustrate the feasibility of our algorithm on a web-based prototype that works with trajectories derived from mouse and touch input. As can be observed, our proposal outperforms the classical K-means algorithm in terms of accuracy (better, well-formed segmentations) and performance (less computation time).
Interaction-based user interface redesign BIBAFull-Text 311-312
  Luis Leiva
This paper demonstrates a general framework to restyle UI widgets, in order to adapt them to the user behavior. Different implementation examples illustrate its feasibility. The value of this methodology comes from the fact that it is suited to any application language or toolkit supporting structured data hierarchies and style sheets; e.g., interfaces created in HTML, XUL, Flex/AIR (ActionScript), or Java. As described in the paper, an explicit end user intervention is not required, and changes are gradually applied so that they are not intrusive for the user.
Explain-a-LOD: using linked open data for interpreting statistics BIBAFull-Text 313-314
  Heiko Paulheim
While statistics are omnipresent, e.g., depicting the corruption in different countries, it is often not trivial to find the explanation for a statistical effect, e.g., why the corruption is higher in some countries than in others. The necessary facts that can explain a statistic are often not contained in the statistics file itself. This demo shows Explain-a-LOD, a tool for generating possible explanations for statistics from Linked Open Data. The tool accepts statistical data as input, and it automatically retrieves data from the Linked Open Data cloud and generates possible explanations.
Smart interface for reshaping photos in 3D BIBAFull-Text 315-316
  Rui Nóbrega; Nuno Correia
In this work we present an interactive prototype for an interface that supports the interaction with virtual objects integrated in a real life scenario. The user can reshape or re-design a real space with virtual objects using several pictures of the desired space. The images are analyzed for known features such as surfaces, edges, floor location and room orientation. Using these elements, it is possible to devise an augmented reality system where the user can add virtual objects to the scenario. The smart interface attaches the objects to the scene elements (e.g., floor) automatically. The current demo loads images from user files or takes a snapshot directly from the camera. The high-level features are automatically detected, but can be manually adjusted. A use-case example of an augmented reality application is presented.
RadSpeech's mobile dialogue system for radiologists BIBAFull-Text 317-318
  Daniel Sonntag; Christian Schulz; Christian Reuschling; Luis Galarraga
With RadSpeech, we aim to build the next generation of intelligent, scalable, and user-friendly semantic search interfaces for the medical imaging domain, based on semantic technologies. Ontology-based knowledge representation is used not only for the image contents, but also for the complex natural language understanding and dialogue management process. This demo shows a speech-based annotation system for radiology images and focuses on a new and effective way to annotate medical image regions with a specific medical, structured, diagnosis while using speech and pointing gestures on the go.
Airwriting: demonstrating mobile text input by 3D-space handwriting BIBAFull-Text 319-320
  Christoph Amma; Tanja Schultz
We demonstrate our airwriting interface for mobile hands-free text entry. The interface enables a user to input text into a computer by writing in the air like on an imaginary blackboard. Hand motion is measured by an accelerometer and a gyroscope attached to the back of the hand and data is sent wirelessly to the processing computer. The system can continuously recognize arbitrary sentences based on a predefined vocabulary in real-time. The recognizer uses Hidden Markov Models (HMM) together with a statistical language model. We achieve a user-independent word error rate of 11% for a 8K vocabulary based on an experiment with nine users.
Wheeling around with Wanteat: exploring mixed social networks in the gastronomy domain BIBAFull-Text 321-322
  Fabrizio Antonelli; Giulia Biamino; Francesca Carmagnola; Federica Cena; Elisa Chiabrando; Luca Console; Vincenzo Cuciti; Matteo Demichelis; Franco Fassio; Fabrizio Franceschi; Roberto Furnari; Cristina Gena; Marina Geymonat; Piercarlo Grimaldi; Pierluigi Grillo; Elena Guercio; Silvia Likavec; Ilaria Lombardi; Dario Mana; Alessandro Marcengo; Michele Mioli; Mario Mirabelli; Monica Perrero; Claudia Picardi; Federica Protti; Amon Rapp; Roberta Sandon; Rossana Simeoni; Daniele Theseider Dupré; Ilaria Torre; Andrea Toso; Fabio Torta; Fabiana Vernero
Wanteat is a framework and a suite of applications which allow users to interact with and explore mixed social networks of smart objects and people in the gastronomy domain, thus promoting the cultural heritage of a territory. Wanteat interaction model is based on the concept of a "wheel" [1].
KMulE: a framework for user-based comparison of recommender algorithms BIBAFull-Text 323-324
  Alan Said; Ernesto William De Luca; Benjamin Kille; Brijnesh Jain; Immo Micus; Sahin Albayrak
Collaborative Filtering Recommender Systems come in a wide variety of variants. In this paper we present a system for visualizing and comparing recommendations provided by different collaborative recommendation algorithms. The system utilizes a set of context-aware, hybrid, and other collaborative filtering solutions in order to generate various recommendations from which its users can pick those corresponding best to their current situation (i.e. context). All user interaction is fed back to the system in order to additionally improve the quality of the recommendations. Additionally, users can explicitly ask the system to treat certain recommenders as more important than others, or disregard them completely if the list of recommended movies is not to their liking.
A prototype for interactive speech transcription balancing error and supervision effort BIBAFull-Text 325-326
  Isaias Sanchez-Cortina; Nicolás Serrano; Alberto Sanchis; Alfons Juan
A system to transcribe speech data is presented following an interactive paradigm in which both, the system produces automatically speech transcriptions and the user is assisted by the system to amend output errors as efficiently as possible. Partially supervised transcriptions with a tolerance error fixed by the user are used to incrementally adapt the underlying system models. The prototype uses a simple yet effective method to find an optimal balance between recognition error and supervision effort.
Seamless online/offline shopping experience design for in-store customers BIBAFull-Text 327-328
  Udana Bandara
Over the last few years, the conventional brick and mortar business model has been challenged by the proliferation of smartphone-based shopping apps, which exploit the weaknesses of this conventional model. As an alternative to these apps, we have developed Ubira [1], a patent-pending service platform that allows healthy online/offline competition rather than merely exploiting the weaknesses. This business model provides brick and mortar shops a fair chance to compete with online stores while creating a seamless shopping experience for in-store customers based on an online/offline partnership. The main design challenge in Ubira has been to promote serendipity in shopping rather than bargain hunting, and integrate the legacy inventory systems of brick and mortar businesses into the platform. To overcome these challenges, we have made some critical design choices based on context awareness and seamful design methods.
icAuth: image-color based authentication system BIBAFull-Text 329-330
  Pramod Verma
Authentication interfaces are GUIs that provide the protection for an application or system. In this paper, we present icAuth: a novel image and color based authentication interface for the authentication process. We enhance the existing Image Based Authentication (IBA) with an additional interactive method. In our approach, the user not only chooses image(s) as a key during the registration process, but also clicks on various regions on the image to generate an additional key. This additional key is in the form of a sequence of colors that correspond to the clicked areas. In essence, the user chooses a color sequence along with the selected images. During the next authentication process, the user has to produce the same color sequence on the recognized images. The user is required to remember the same switching sequence among the images, without having to memorize the precise location of the initial clicks during setup.
A web-based user interface for interaction with hierarchically structured events BIBAFull-Text 331-332
  Sven Buschbeck; Anthony Jameson; Tanja Schneeberger; Robin Woll
Intelligent technologies have been used in various ways to support more effective representation and processing of media and documents in terms of the events that they refer to. This demo presents some innovations that have been introduced in a web-based interface to a repository of media and documents that are organized in terms of hierarchically structured events.
PINTER: interactive storytelling with physiological input BIBAFull-Text 333-334
  Stephen Gilroy; Julie Porteous; Fred Charles; Marc Cavazza
The dominant interaction paradigm in Interactive Storytelling (IS) systems so far has been active interventions by the user by means of a variety of modalities. PINTER is an IS system that uses physiological inputs -- surface electromyography (EMG) and galvanic skin response (GSR) [1] -- as a form of passive interaction, opening up the possibility of the use of traditional filmic techniques [2, 3] to implement IS without requiring immersion-breaking interactive responses. The goal of this demonstration is to illustrate the ways in which passive interaction combined with filmic visualisation, dialogue and music, and a plan-based narrative generation approach can form a new basis for an adaptive interactive narrative.
A flexible platform for developing context-aware 3D gesture-based interfaces BIBAFull-Text 335-336
  Ali Bigdelou; Loren Schwarz; Tobias Benz; Nassir Navab
In this paper, we introduce a flexible framework that can facilitate the definition of 3D gesture-based interfaces. Highlighting the need for context awareness in complex domains, such as the operating room, we argue how the proposed architecture can overcome integration challenges. Through a real-life scenario, an intra-operative medical image viewer, we demonstrate how the proposed framework can be used in practice to define user interfaces in collaborative environments, where the behavior and the system response can be adapted based on the current workflow stage and individual user requirements. Finally, we demonstrate how the defined interface can be manipulated using a high-level visual programming interface. The extensibility of the proposed architecture makes it applicable to a wide range of scenarios.
First demonstration of the intelligent reviewer's assistant BIBAFull-Text 337-338
  Ruihai Dong; Kevin McCarthy; Michael O'Mahony; Markus Schaal; Barry Smyth
User opinions and reviews are an important part of the modern web and all major e-commerce sites typically provide their users with the ability to provide and access customer reviews across their product catalog. The importance of reviews has driven the need to improve the review quality by providing interactive support for the reviewer and we will demonstrate the first version of an Intelligent Reviewer's Assistant for this purpose. Our browser plugin is designed to work with major sites like Amazon and to provide users with suggestions as they write their reviews. In particular, these suggestions take the form of topics (e.g. product features) that a reviewer may wish to write about and the suggestions automatically adapt as the user writes their review.
Collecting multimodal data in the wild BIBAFull-Text 339-340
  Michael Johnston; Patrick Ehlen
Multimodal interaction allows users to specify commands using combinations of inputs from multiple different modalities. For example, in a local search application, a user might say "gas stations" while simultaneously tracing a route on a touchscreen display. In this demonstration, we describe the extension of our cloud-based speech recognition architecture to a Multimodal Semantic Interpretation System (MSIS) that supports processing of multimodal inputs streamed over HTTP. We illustrate the capabilities of the framework using Speak4itSM, a deployed mobile local search application supporting combined speech and gesture input. We provide interactive demonstrations of Speak4it on the iPhone and iPad and explain the challenges of supporting true multimodal interaction in a deployed mobile service.

Doctoral consortium

Deducing parameters for personalizing maps from map interaction patterns BIBAFull-Text 341-344
  Beate Weninger
Interactive maps on the internet have become frequently used means to convey spatial information to the public. However, many maps are not developed to suit a variety of users and thus lead to frustration. To user-center maps we therefore recommend to personalize them to individual users. As many parameters that can be used as a trigger for personalization are not easy to be logged on the internet, we suggest user-map interaction. Interaction can be easily tracked and gives comprehensive information about map use. Since no interpretation of user-map interaction is available it is the aim of this PhD to observe interaction, and to evaluate and interpret it. We hypothesize that there are map interaction patterns, means recurring sequences of consecutive actions which are necessary to complete a task. Our goal is to deduce parameters for personalization from these map interaction patterns.
Evaluating an organic interface for learning mathematics BIBAFull-Text 345-348
  Bee Suan Wong
The current formats used for presenting mathematics either on paper or in electronic form have usability limitations that make learning mathematics challenging. The concept of an Organic User Interface, promises a natural interface that blends with the human ecology system and therefore affords smoother transition and improved usability. This research aims to examine how the affordances of an Organic User Interface influence users learning of important mathematical concepts. The relationship between learning time and the usability factors, or affordances of an Organic User Interface will be determined and contrasted with those of Graphical User Interfaces.
Generation Y interactions BIBAFull-Text 349-352
  Wei Liu
Information technology (IT) support of office work has increased rapidly in functionality, but the interaction styles have evolved more slowly. This project explores interaction design opportunities of IT supported tools in the context of office work. A series of (contextual) interviews was conducted with Generation Y office workers, aiming to identify their interaction qualities. Three interactive prototypes were built to map these interaction qualities and to demonstrate future ways of working. This project resulted in a set of design guidelines, aiming to support Generation Y interactions in future office work. Designers and researchers who focus on understanding (rich interactions in) the work context would benefit from the result of this project.
Active forms for responsive environments BIBAFull-Text 353-358
  Simon Scott-Harden
Active Forms are interactive artefacts that are a focal point of attention for the user. Such devices are interactive and can change shape. They are embedded with both sensors and actuators and are a visualisation and embodiment of some application or service. To be more specific, Active Forms are defined as interactive products or devices that can render content thanks to perceptible changes to their physical form and appearance. We see Active Forms as ideal gateway for the interaction with and the control of Responsive Environments (RE) as defined in [1].
   Tangible interaction is the precursor of Active Forms, it was essentially about coupling digital content and physical elements of an interface in an integrating combination. The content is about the internal state of the products and about some of the application(s) and service(s) they support. The tangible interaction focuses on the interface or systems that are physically embodied in the physical artifact. Tangible User Interfaces (TUI) are reactive devices that require a user input to change shape. Active Forms, on the other hand are interactive devices that change shape and appearance. The changes in the Active Forms are a result of either, user actions or internal actuators, both the physical form, such as shape or size, and the appearance, such as colour or temperature, can change.
   Within Active Forms, there is a balance to be had between the cognitive load on the user, the selection of modalities, the media bandwidth and the user attention. The aim being to have the Active Forms as the user focus of interaction and attention. We have listed below some of the key features of Active Forms:
  • Active Forms are interactive devices that are both reactive to user actions
       and proactive in displaying information.
  • Active Forms are a gateway to applications or services within the RE and
       there is a change to the internal state of the device.
  • The user actions and the device reactions of an Active Form are merged and
       are spatially co-located.
  • Active Forms act as their embodiment, as physical objects, Active Forms also
       have aesthetic value per se. The changes in the Active Forms are a result of either, user actions or internal actuators, both the physical form, such as shape or size, and the appearance, such as colour or temperature, can change.
  • Using theories of intrinsic motivation to support ICT learning for the ageing population BIBAFull-Text 359-362
      Tanya Goldhaber
    Access to Information and Communication Technology (ICT) has the potential to improve the quality of life for many members of the ageing population. However, some older users lack the intrinsic motivation to learn to use this technology, and poor user interface design is partly to blame. The research presented here investigates how motivation theory can be applied to interface design in order to encourage older users to learn to use ICT. In addition to a brief literature review, an overview of methods, research goals, and current research status are presented.
    Virtual marionette BIBAFull-Text 363-366
      Luís Leite
    Virtual Marionette is a research on digital puppetry, an interdisciplinary approach that brings the art of puppetry into the world of digital animation. Inspired in the traditional marionette technology our intention is to study novel interfaces as an interaction platform for creating artistic contents based on computer animated puppets. The overall goal of this thesis is to research and deploy techniques and methods for the manipulation of articulated puppets in real-time with low-cost interfaces to establish an interaction model for digital puppetry.
    Eliciting evaluative comments from users in web 2.0 scenarios BIBAFull-Text 367-370
      Richard Gomer
    In recent years, we've seen a huge growth in the level of user-supplied reviews posted online. These reviews range from feedback on eBay or comments on sites such as YouTube, to social bookmarking sites like StumbleUpon that allow users to comment on almost any page on the web.
       I'm interested in how these comments are incorporated into evaluative judgements by the users that read them, and how we can improve them through better user interfaces in order to maximise their value to other users. The work draws on psychology and neurology, as well as ideas around credibility from information science, to design and test the impact of intelligent interface changes and behaviour on review composition and ascertain how the composition of a review can make it more or less useful.
    Graph based user modeling BIBAFull-Text 371-374
      Amit Tiroshi
    An overload in service applications and websites induced by ubiquitous connectivity has brought the need for personalization, as a way to cope with it. However, the need for providing every service with a user model calls for interoperable user models, since user details are scattered in many different systems (e.g., online services such as mail/banking/healthcare/ecommerce sites/social networks), each storing the user's details, preferences and history in different representations and data formats. Various approaches for user modeling interoperability were studied from different perspectives (general ontologies, personalized ontologies, mediation), but so far the challenges are yet to be met. This paper proposes a new way for representing user models, an abstracted graph based one, which will support both interoperability and advanced user modeling features.
    An explanation-centric approach for personalizing intelligent agents BIBAFull-Text 375-378
      Todd Kulesza
    Intelligent agents are becoming ubiquitous in the lives of users, but the research community has only recently begun to study how people establish trust in and communicate with such agents. I plan to design an explanation-centric approach to support end users in personalizing their intelligent agents and in assessing their strengths and weaknesses. My goal is to define an approach that helps people understand when they can rely on their intelligent agents' decisions, and allows them to directly debug their agents' reasoning when it does not align with their own.
    Design for reflection on health behavior change BIBAFull-Text 379-382
      Feng Gao
    Although the persuasion-based health behavior change systems have achieved certain success particularly in motivating physical activity, researchers now start criticizing that persuasion-based systems have problems in taking over too much control, paying not enough attention to people's thinking, and failing in acknowledging external constraints and exploring resources. The alternative notion of reflection has been supported by different researchers' views, and in my thesis work I aim to explore this notion in the context of dietary change. The main goal of my thesis work is to explore what people think in the different food-related activities and apply those understandings into system designs to foster and assist people's reflection on everyday dietary change.
    Managing implicit assumptions in natural language interfaces BIBAFull-Text 383-388
      Dustin Smith
    A person's choice of what to communicate and how to communicate it depends on the information he or she believes is shared with the audience. This presents a challenge for natural language interfaces, because it is hard for people to predict what information they share with the interface and how it will use this information to interpret their text. This is especially difficult for pragmatic-level assumptions supplied by the interpreter that go beyond the information in the surface text, because these assumptions are negotiated in dialogue and frequently revised or redacted.
       We have built a calendaring interface that allows users to communicate English event descriptions. This constrained task gives us a clear criteria for communication success and failure. Failures are opportunities to acquire and revise assumptions: to collect lexical and semantic knowledge from a variety of users. By lowering the interaction barrier so end users can contribute to the linguistic interpretation process, we can collect culture-specific lexical and semantic knowledge directly from the members of the cultural group who possess it. This knowledge is essential for the pragmatic task of deriving what a speaker meant from what they said.
       The goal of this research is to make the assumptions involved with interpreting natural language explicit to the user. Using a model of language generation and interpretation based on planning and plan recognition, we capture, through user contributions, word definitions and commonsense assumptions -- and we represent both as belief-changing actions. Using visualizations and a direct manipulation interface, users can access the interpretation status, inspect which assumptions were made, and suggest or modify existing assumptions. With the aim of providing the functionally equivalent of the negotiation stage in interpersonal dialogue, we evaluate the interface by how it allows users to revise and extend assumptions toward successful interpretation.
    Quantitative analysis and inference on gaze data using natural language processing techniques BIBAFull-Text 389-392
      Pascual Martinez-Gomez
    Eye-tracking devices find applications in human-machine interaction, hypothesis testing in psycholinguistic and usability studies, relevant feature extraction when designing models related to human behavior and to build user-centered information systems. We aim at providing a general and robust framework to do quantitative analysis and inference using data collected by eye-trackers when users read text. To achieve this objective, first the accuracy of eye-trackers has to be increased beyond sensor capabilities by using information from the content or the structure of the text. Then, natural language processing techniques will be used to process text appearing on the screen and the recognized reading word sequence. Within this framework, it will be possible to better understand user's intentions, record knowledge acquisition and predict information needs. The intention is to build a user model and user model of the World from texts that users have read. This opens the door to more personalized systems with on-line adaptation capabilities.


    2nd workshop on location awareness for mixed and dual reality (LaMDa'12) BIBAFull-Text 393-396
      Tim Schwartz; Gerrit Kahl; Teemu Pulkkinen; Petteri Nurmi; Eyal Dim; Sally Applin
    The workshop explores the interactions between location awareness and Dual/Mixed Reality in smart environments and the impact on culture and society. The main scope of this workshop is: How can the Dual Reality paradigm be used to improve applications in smart environments and which new possibilities are opened up by these paradigms? This includes positioning methods and location-based services using the DR paradigm, such as navigation services and group interaction services (location-based social signal processing). The workshop is also open to discuss sensor and actuator technologies that may help to realize the synchronization of the virtual and real world.
       The main scope of this workshop is: How can the Dual Reality paradigm be used to improve location-based and socially-aware services and other applications in smart environments?
    1st international workshop on user modeling from social media BIBAFull-Text 397-400
      Jalal Mahmud; Jeffrey Nichols; Michelle Zhou
    Massive amounts of data are being generated on social media sites, such as Twitter and Facebook. People from all walks of life share data about social events, express opinions, discuss their interests, publicize businesses, recommend products, and, explicitly or implicitly, reveal personal information. This workshop will focus on the use of social media data for creating models of individual users from the content that they publish. Deeper understanding of user behavior and associated attributes can benefit a wide range of intelligent applications, such as social recommender systems and expert finders, as well as provide the foundation in support of novel user interfaces (e.g., actively engaging the crowd in mixed-initiative question-answering systems). These applications and interfaces may offer significant benefits to users across a wide variety of domains, such as retail, government, healthcare and education. User modeling from public social media data may also reveal information that users would prefer to keep private. Such concerns are particularly important because individuals do not have complete control over the information they share about themselves. For example, friends of a user may inadvertently divulge private information about that user in their own posts. In this workshop we will also discuss possible mechanisms that users might employ to monitor what information has been revealed about themselves on social media and obfuscate any sensitive information that has been accidentally revealed.
    3rd workshop on semantic models for adaptive interactive systems (SEMAIS): (SEMAIS) BIBAFull-Text 401-404
      Tim Hussein; Stephan Lukosch; Heiko Paulheim; Jürgen Ziegler; Gaëlle Calvary
    The International Workshop on Semantic Models for Adaptive Interactive Systems (SEMAIS 2012) aims to identify emerging trends in interactive system design using semantic models.
    Developing intelligent user interfaces for e-accessibility and e-inclusion BIBAFull-Text 405-408
      Pradipta Biswas; Pat Langdon; Christoph Jung; Pascal Hamisu; Carlos Duarte; Luis Almeida
    This workshop aims to gap the bridge between mainstream research on intelligent systems and accessibility researchers by presenting papers and demonstrations on developing adaptable multimodal systems for elderly and disabled users. The workshop is organized in the context of EU GUIDE project and focus on Web and Digital TV applications. However the research and applications are relevant for different platforms like computers, tablet and ubiquitous devices. The workshop consists of a keynote speech on standardization of developing intelligent and accessible system followed by five paper and demonstration presentations. A set of papers from this workshop will later appear at the International Journal of Digital Television.
    2nd Workshop on context-awareness in retrieval and recommendation: (CaRR 2012) BIBAFull-Text 409-412
      Ernesto William De Luca; Matthias Böhmer; Alan Said; Ed Chi
    Context-aware information is widely available in various ways and is becoming more and more important for enhancing retrieval performance and recommendation results. The current main issue to cope with is not only recommending or retrieving the most relevant items and content, but defining them ad hoc. Other relevant issues include personalizing and adapting the information and the way it is displayed to the user's current situation and interests. Ubiquitous computing further provides new means for capturing user feedback on items and providing information.