IDInteraction: Capturing Indicative Usage Models in Software for Implicit Device Interaction

Lead Research Organisation: University of Manchester
Department Name: Computer Science

Abstract

The IDInteraction project asks: can we exploit models of human behaviour to move away from direct, unambiguous user commands, towards seamless user-device interaction? It will investigate and develop the techniques to capture 'Indicative Usage Models (IUMs), behavioural patterns or cues that precede a particular activity, and translate these into software-based 'Indicative Usage Patterns' (IUPs), to drive interaction with an app. The project will focus on future television broadcasting, examining the extent to which it is possible to capture IUMs from device sensor and event data, and deploy these as IUPs to pull content (additional information or related activities) to a 'second screen companion app', which a viewer watches on a mobile device alongside a TV programme.

The core of the research - investigating the extent to which we can anticipate user commands and requests - will play an important role in shaping the way we use future technology. The Internet of Things is fast becoming a reality, but the appropriate paradigms for interacting with it are far from understood. A rapidly growing number of connected devices means a potentially vast increase in the types of technology which people must use successfully. Whilst a person may be prepared to learn how to use a new home automation system or car, expecting him or her to do this for every machine with which there is a cursory encounter is unrealistic. We have already started to understand that interactive experiences which take fundamental human perception and thought processes into account are more successful than those which require significant learning on the part of the user. Implicit Device Interaction, where devices automatically know what the user wants, before a command is issued, is the next step.

The scenario investigated - second screen viewing - is also particularly timely. Broadcasters are keen to exploit the creative potential of a context where people are watching television with a mobile device, not least because it is set to become the principle form of TV viewing in the next few years. At present, however, the scenario is not well understood: whilst some research has focused on social aspects of this situation, very little has examined the perceptual or behavioural aspects of second screen interaction. Current companion apps, designed to complement the main programme, push information to the mobile device at given points in time, and this change on the secondary screen, which occurs in peripheral vision, is potentially distracting. A situation where information flow intuitively stops and starts according to the location of the viewer's attention is a highly desirable goal.

IDInteraction is an ambitious research project, investigating an aspect of human behaviour that is crucial to the future development of implicit user interfaces, and has been planned to tackle the problem from end-to-end. By studying a slice of the research problem, from modelling behaviour to testing a new user interface, it will provide an overview of the theoretical and technical challenges that the development of seamless user-device interaction will entail, and flag key areas for further investigation. From the perspective of effective software development, the project entails a considerable degree of risk: IUMs may be difficult to capture and deploy, and seamless information provision may be challenging to implement in this context. From the perspective of building theory in this area, such results would still have significant value, however. Improving our understanding of the limits of implicit interaction is crucial to moving this important, emergent, field forward.

Planned Impact

IDInteraction will create societal and economic impacts from two perspectives.

(1) Determining the potential of using device sensor and event data to anticipate and pre-empt user commands and requests, leading to a significant shift, and potentially a major improvement, in the way in which we interact with technology.

(2) Investigating the increasingly common but under-researched home entertainment activity of watching television with a 'second screen' mobile device, and offering a rich understanding of the ways in which the viewing experience can be improved, in particular through the seamless provision of information at the moment the user requires it.

The impact of (1) is potentially very wide; the impact of (2) will be felt primarily within the domain of television broadcasting.

Key impacts, as described in the RCUK Typology, are therefore as follows:

'Contributing towards wealth creation and economic development'
IDInteraction will impact on how we use devices in the future. The models developed will be based, in the first instance, on interactivity during second screen viewing, but the techniques used for both creating Indicative Usage Models, and mapping them to software-based Indicative Usage Patterns, will have potential relevance for any interactive device. In particular, IDInteraction will contribute to our understanding of how to develop successful user interfaces for the rapidly growing 'Internet of Things', through examining anticipatory interaction with multiple networked devices. The step-change in interaction promised by this research will enhance 'economic prosperity' in the UK by improving the user experience of interactive devices, and enabling people to access information or services more quickly (even half a second saved during an interaction will have a significant effect when considered over long periods of time and across multiple users).

'Shaping and enhancing the effectiveness of a Public Service Broadcaster.'
The models and software outputs that will result from the project have already been identified by the BBC to be of direct relevance to their mission of delivering distinctive and transformational services both to the UK and internationally. The systematic investigation of second screen viewing entailed by IDInteraction and the resulting ability to provide content precisely when the viewer requires it, will both improve the user experience of watching television, and open the door to gathering a more detailed understanding of how people watch television with a mobile device.

'Enhancing the research capacity, knowledge and skills of a Public Service Broadcaster.'
The Department for Culture, Media and Sport have recently identified the production of 'world-beating, innovative content and services that originate here in the UK, but that are in demand across the globe,' as a strategic priority. UK media organisations, who have always been world leaders in broadcast technology, are finding it increasingly challenging to compete on the global stage (Future of Innovation in Television Technology Task Force Report, 2014). This research is therefore particularly timely from the perspective of supporting the UK television industry.

'Enhancing cultural enrichment and quality of life.'
The improvement to the viewing experience offered by the user-focused novel interaction paradigms investigated in IDInteration will also contribute to the Digital Economy 'Communities and Culture' theme. Watching television whilst using a mobile device is considered to be the dominant viewing environment of the future, and supporting it as effectively as possible will enhance education, enjoyment, relaxation and well-being.

Publications

10 25 50
publication icon
Apaolaza A (2016) ABC

publication icon
Brown A (2019) Contrasting delivery modes for second screen TV content-Push or pull? in International Journal of Human-Computer Studies

 
Description We have demonstrated that it is possible to use object tracking to 'code' interaction behaviour (in this case, determine the location of someone's visual attention). This is significant because behavioural coding is usually done manually, by watching videos in slow motion, and is therefore extremely time-consuming, subjective and hard to quantify. Of particular note is that training with video frames selected at random, rather than sequentially, results in a massive improvement in performance and efficiency.
Exploitation Route We have developed open source object tracking and behavioural recognition software that can be used by other researchers, and the BBC.
Sectors Creative Economy,Digital/Communication/Information Technologies (including Software)

URL http://idinteraction.github.io
 
Description The findings with regard to training machine learning algorithms were presented in a BCS/IET talk in 2019, garnering interest from the public and industry (IBM) with regard to its potential impact, and will be presented at a University of Birmingham seminar in 2022. This has also resulted in a societal benefit from the perspective of the public understanding of science.
Sector Digital/Communication/Information Technologies (including Software),Education
Impact Types Cultural,Societal

 
Title IDInteraction Object Tracking Docker Image 
Description Takes a video and bounding box position as input, performs tracking on the object within the bounding box, and outputs a video with the bounding box overlaid, and a CSV file containing the positional and rotational information for the object. 
Type Of Material Improvements to research infrastructure 
Year Produced 2015 
Provided To Others? Yes  
Impact Execution environment used for the research described in http://dx.doi.org/10.1145/2851581.2892483. 
URL https://hub.docker.com/r/idinteraction/object-tracking/
 
Title OpenFace Docker Image 
Description Takes a directory of videos and outputs facial tracking data for each frame of each video. 
Type Of Material Improvements to research infrastructure 
Provided To Others? No  
Impact Reproducible execution environment for generating facial tracking data 
URL https://hub.docker.com/r/idinteraction/openface/
 
Description BBC Data Science Research Partnership 
Organisation British Broadcasting Corporation (BBC)
Department BBC Research & Development
Country United Kingdom 
Sector Public 
PI Contribution The relationship developed during IDInteraction formed the basis for the University of Manchester forming an official Data Science Research Partnership (DSRP) with the BBC and seven other universities. The Partnership is led by Jay at Manchester. IDInteraction analysis techniques are currently being applied in Jonathan Carlton's iCASE studentship, and Joshua Woodcock's internship, but it should be noted that the reach of the DSRP extends far beyond this, touching every faculty in the University.
Collaborator Contribution The DSRP has led to seven directly funded three month internships, three iCASE PhD studentships, and an ESRC IAA. In addition to the direct funding, the BBC is contributing access to its unique datasets, on which it is very difficult to place an accurate in kind value.
Impact The projects have only started over the last few weeks, and there are thus no concrete outputs so far.
Start Year 2017
 
Title CppMT (IDInteraction version) 
Description This is a modification to the CppMT object tracking tool, to output the object tracking information (position and rotation of object) in CSV format, and to allow more accurate definition of the bounding box. 
Type Of Technology Software 
Year Produced 2015 
Open Source License? Yes  
Impact This has enabled object tracking information to be used for behavioural modelling, which has thus far been applied to determining the location of attention of people watching television with a mobile device (see http://dx.doi.org/10.1145/2851581.2892483). Modifications have also been merged into the original CppMT tool by the original author, and are available as open source software for use and reuse by other researchers. 
URL https://github.com/IDInteraction/CppMT
 
Title CppMT-replay 
Description CppMT-Replay takes a video and associated object tracking data output from CppMT (https://github.com/gnebehay/CppMT) and combines the two, overlaying the bounding box outline and centre point in the output video. 
Type Of Technology Software 
Year Produced 2015 
Open Source License? Yes  
Impact This software enables users to review the outputs of the object tracking of the CppMT tool by eye, to determine its accuracy. It has been used to validate the results of behavioural coding using CppMT (IDInteraction Version) by eye (see http://dx.doi.org/10.1145/2851581.2892483), as a precursor to quantitative analysis. 
URL https://github.com/IDInteraction/CppMT-replay
 
Title abc-display-tool 
Description abc-display-tool allows users to classify behaviours in a random subset of video frames. External tracking data, from e.g., CppMT (https://github.com/gnebehay/CppMT) is used to construct a Decision Tree Classifier, that is used to predict behaviours in the remaining frames. The accuracy of the classifier can be evaluated using external ground-truth data, or via a cross-validation approach. 
Type Of Technology Software 
Year Produced 2017 
Open Source License? Yes  
Impact The tool has the potential to dramatically reduce the time needed to classify behaviors in video streams. 
URL https://github.com/IDInteraction/abc-display-tool
 
Description Data science conference poster 
Form Of Engagement Activity Participation in an activity, workshop or similar
Part Of Official Scheme? No
Geographic Reach International
Primary Audience Professional Practitioners
Results and Impact Presented a poster on our research; "Training a Tool to Recognise Behaviour From Video"
Year(s) Of Engagement Activity 2017
 
Description Docker workshop 
Form Of Engagement Activity A talk or presentation
Part Of Official Scheme? No
Geographic Reach International
Primary Audience Other audiences
Results and Impact Presentation of our approach to using Docker to improve the reproducibility of the research process
Year(s) Of Engagement Activity 2017
URL http://idinteraction.cs.manchester.ac.uk/dockerworkshop/DockerWorkshop.html#/
 
Description Research Software Engineers Conference talk 
Form Of Engagement Activity A talk or presentation
Part Of Official Scheme? No
Geographic Reach International
Primary Audience Professional Practitioners
Results and Impact Presented our approach to treating reproducible research as a software engineering issue. This sparked questions and an interesting discussion afterwards; people were interested in the approach we'd taken.
Year(s) Of Engagement Activity 2017
URL http://idinteraction.cs.manchester.ac.uk/RSE2017Talk/ReproducibleResearchIsRSE.html#/
 
Description Turing Insight Talk Manchester 
Form Of Engagement Activity A talk or presentation
Part Of Official Scheme? No
Geographic Reach National
Primary Audience Public/other audiences
Results and Impact One of the project outcomes with regard to training machine learning algorithms was the focus for the annual Manchester BCS/IET Turing Insight talk:

Spot the Difference: What can machine learning tell us about humans?

Recognising simple behaviours - pointing, sitting, nodding - is easy for humans. The ability of machines to do this is advancing rapidly, but even the most sophisticated systems must be 'trained' with a large number of examples to achieve a reasonable level of accuracy. This talk describes what we can learn from teaching a machine to recognise when a person is watching television, and when they are looking at a mobile phone. The results reveal the differences between the way humans learn, and the way machines do, and show how the assumptions people make when building technology can strongly bias the resulting system behaviour.
Year(s) Of Engagement Activity 2019
URL https://www.bcs.org/events-home/turing-talk/2019-turing-talk/insight-speakers/