![]() |
Principles for Improving Interaction in Telephone-Based InterfacesEPSRC project GR/L66373Oct/97 - Oct/2000Project abstractTelephone-based interfaces (TBIs) are an increasingly important method of interacting with computers systems (such as electronic banking and voicemail). Telephones themselves are also incorporating greater functionality (such as address books and call forwarding). In both cases this extra functionality may be rendered useless if usablity is not considered. One common usability problem is users getting lost when navigating through hierarchies of options of functions. This may mean that some functions are not used or that users cannot the goals they wish. The innovative aspect of this proposal is to use structured non-speech sounds (such as short pieces of music) to enhance the output of information in TBIs. Sound can present information rapidly without getting in the way of any speech output. I will investigate the use of sound to provide navigation cues to stop users getting lost and also to provide richer output methods to create more flexible interaction techniques. To ensure effectiveness i will perform full usability evalustion. TBI designers will benefit from this research because guidelines produced will enable them to create more powerful interfaces. End users will benefit because the resulting telephones and telephon services will be more usable. ResearchersPrincipal investigator was Prof Stephen Brewster. Gregory Leplatre was employed on the project as a research student. Seppo Helle from Nokia was involved in the project as an industrial collaborator. Earcons and telephone-based interfacesWe have done several experiments to investigate the use of earcons for representing hierarchical information. This has applications for telephone-based interfaces (such as phone-banking or voicemail), the design of mobile phones and blind computer users. Telephone-based interfaces (TBIs) are becoming an increasingly important method for interacting with computer systems. The telephone is an ubiquitous device and is many people’s primary method of entry into the information infrastructure. Access to an increasing number of services is being offered over the telephone, such as voice-mail, electronic banking and even Web pages. The rapidly increasing use of mobile telephones means that people access these services at many different times and places. Telephones themselves are now also incorporating greater functionality (such as multi-party calling, address books, diaries or call forwarding). The provision of this extra functionality may be rendered useless if usability issues are not considered. Little use has been made of structured non-speech sounds in TBIs. Our work has shown that the use of such sounds can increase the bandwidth of communication between the system and the user, allowing a richer interaction. These ‘multimedia’ telephone interfaces are more usable than their current eqivalents. Sound has many advantages. For example, it is good for communicating information quickly. Unlike speech, non-speech sound is universal; the user is not tied to one language, which is important for the increased international use of computer systems. There is also great potential for the results of this work in other non-graphical interfaces such as those for visually disabled people and those where working conditions or protective clothing mean that a screen cannot be used. The projectThese pages will tell you more about this general area of research (also have a look at Gregory Leplatre's web pages). There is a hypercard stack full of earcons to download too. We have several more sets of earcons that we can supply if you contact us. There is an initial java demo of how temporal cues can be used for navigation (this made up experiment 2 of our project). We then developed a detailed Java simulation of how sounds might be used for navigation in a real mobile phone (which formed experiment 3 of our project). This can be downloaded - please let us know if you use it. To enable interface designers who are not sound experts to develop sonically-enhanced interfaces we have developed a sophisticated tool to allow the creation of hierarchical earcons. You can download this - again let us know if you use it. In the final stages of the project we investigated soundgraphs to present complex, continous data. We looked at presenting stock market data on a mobile telephone. The value of the stock was mapped to the pitch of the note, so that users heard a change in pitch over time. Full details can be found in the paper published. If you are interested we can give you the Java application we used to test our ideas. PublicationsThe main publications in this area are shown below, the numbered items were produced as part of the project (the group publications list has the most up to date complete list of papers):
See our research pages for other projects going on within the Multimodal Interaction Group. |
Prof Stephen Brewster Department of Computing Science University of Glasgow Glasgow, G12 8QQ, UK Tel: +44 (0)141 330 4966 Fax: +44 (0)141 330 4913 Email: stephen@dcs.gla.ac.uk Last Modified: 29 January 2003 |