Ubiquitous Interaction Devices
I investigated how to interface the new emerging class of smartphone devices (circa 2002) with concepts from ubiquitous computing such as location-aware interfaces or context-aware computing. I discovered the surprisingly positive benefits of piggybacking on simple communications medium such as audible sound and visual tags. Our implementations of some of these ended up with new audio ringtone and visual smart tags that worked on the widely deployed mobile phones of the era.
In 2003, the mobile phone market had grown tremendously and given the average consumer access to cheap, small, low-powered and constantly networked devices that they could reliably carry around. Similarly, laptop computers and PDAs became a common accessory for businesses to equip their employees with when on the move. The research question then, was how to effectively interface them with existing digital infrastructure and realise some of the concepts of ubiquitous computing such as location-aware interfaces or context-aware computing.
Ubiquitous Interaction Devices (see original webpages)
was a project I started with
Audio Networking
The project kicked off when we randomly experimented with our fancy Nokia smartphones
and discovered that they didn't have anti-aliasing filters on the microphones.
This allowed us to record and decode ultrasound between the phones. The
2003
In a nutshell, audio networking used ubiquitously available sound hardware (i.e, speakers, soundcards and microphones) for low-bandwidth wireless networking. It has a number of characteristics which differentiate it from existing wireless technologies:
- fine-grained control over the range of transmission by adjusting the volume.
- walls of buildings are typically designed to attenuate sound waves, making it easy to contain transmission to a single room (unlike, for example, Wifi or Bluetooth).
- existing devices can play or record audio to be networked easily (e.g. voice recorders).
The Audionotes video below demonstrates an Audio Pick and Drop Interface: an Internet URL is recorded into a PDA via a standard computer speaker. The URL is later retrieved from the PDA by playing it into the computer, and also printed by carrying the PDA to a print server and playing back the recorded identifier. A great advantage of this is that any cheap voice recorder is capable of carrying the audio information (e.g. mobile phone, PDA, dictaphone, etc).
One of the more interesting discoveries during this research is that most computer soundcards do not have good anti-aliasing filters fitted, resulting in them being able to send and receive inaudible sound at upto 24 Khz (assuming a 48 KHz soundcard). We used this to create a set of inaudible location beacons that would allow laptop computers to simply listen using their microphones and discover their current location without any advanced equipment being required! The location video below demonstrates this.
We then devised a scheme of encoding data into mobile phone ringtones while still making them sound pleasant to the ear. This allows us to use the phone network for the video example above: an SMS 'capability' is transmitted to a phone, which can be played back to gain access to a building. Since the data is just a melody, this allows for uses such as parents purchasing cinema tickets online for their children (who dont own a credit card), yet still allowing the children to gain access to the cinema by playing the ringtone capability back (via their mobile phones, common among children).
We observed that audio networking allows URLs to be easily sent over a telephone conversation, and retrieved over a normal high bandwidth connection by the recipient.
This gets over the real awkwardness of "locating" someone on the Internet, which is rife with restrictive firewalls and Network Address Translation which is destroying end-to-end connectivity. In contrast, it is trivial to locate someone on the phone network given their telephone number, but harder to perform high bandwidth data transfers (which the Internet excels at). There's more detail on these usecases in
SpotCode Interfaces
Once we'd had such success with audible networking, a natural next step was to use the new fancy cameras present on smartphones.
The phone app we built could detect our circular barcodes in realtime, unlike the ungainly "click and wait" experiences of the time. Since the phone computes the relative rotation angle of detected tags in real-time, an intuitive "volume control" was possible by simple locking onto a tag and physically rotating the phone. The videos demonstrates a volume control in the "SpotCode Jukebox", and how further interfaces could be "thrown" to the phone for detailed interactions.
We then built some more elaborate public applications (such as a travel booking shop) in
It also became obvious that the technology was also really robust, since it worked fine on printed (and crumpled) pieces of paper, making it ideal for public signage. We used this to experiment with more robust device discovery in
Towards smarter environments
At the Computer Laboratory, we also happen to have one of the world's best indoor location systems (the Active Bat), which we inherited from AT&T Research when that shut down in Cambridge.
A few really fun projects resulted from this:
- We used the BAT system to help with Bluetooth-based indoor location inference, in
A Study of Bluetooth Propagation Using Accurate Indoor Location Mapping . - Interfaced the audio networking with the AT&T "broadband phones" in
The Broadband Phone Network: Experiences with Context-Aware Telephony . - We constructed a real-time capture the flag game in the Computer Lab with augmented reality "windows" into the game in
Feedback, latency, accuracy: exploring tradeoffs in location-aware gaming . - I worked with my Recoil-in-chief
Nick Ludlam on digital TV interfaces inUbiquitious Computing needs to catch up with Ubiquitous Media .
In around 2005, we sold High Energy Magic Ltd. to a Dutch entrepreneur so that
Activity
About – Research note (Feb 2025)
Using visual tags to bypass Bluetooth device discovery – Note about Using visual tags to bypass Bluetooth device discovery (Jan 2005)
Connecting Paper and Online Worlds by Cellphone – Research note (Oct 2004)
MIT Technology review covers SpotCodes – Research note (Sep 2004)
Cellphone as a virtual mouse/keyboard – Research note (May 2004)
Context-Aware Computing with Sound – Note about Context-Aware Computing with Sound (Oct 2003)
Xen Hypervisor – Project (2002–2009)
References
Interacting with mobile services: an evaluation of camera-phones and visual tags
Eleanor Toye Scott, Richard Sharp, Anil Madhavapeddy, Dave Scott, Eben Upton, and Alan Blackwell.
Journal paper in Personal and Ubiquitous Computing (vol 11 issue 2).
A Study of Bluetooth Propagation Using Accurate Indoor Location Mapping
Anil Madhavapeddy, and Alastair Tse.
Paper in the ubiComp 2005: Ubiquitous Computing.
Ubiquitious Computing needs to catch up with Ubiquitous Media
Richard Sharp, and .
Journal paper in IEEE Pervasive Computing (vol 4 issue 3).
Audio networking: the forgotten wireless technology
Anil Madhavapeddy, Richard Sharp, Dave Scott, and Alastair Tse.
Journal paper in IEEE Pervasive Computing (vol 4 issue 3).
Using smart phones to access site-specific services
Eleanor Toye Scott, Richard Sharp, Anil Madhavapeddy, and Dave Scott.
Journal paper in IEEE Pervasive Computing (vol 4 issue 2).
Using visual tags to bypass Bluetooth device discovery
Dave Scott, Richard Sharp, Anil Madhavapeddy, and Eben Upton.
Journal paper in SIGMOBILE Mob. Comput. Commun. Rev. (vol 9 issue 1).
The Broadband Phone Network: Experiences with Context-Aware Telephony
Ripduman Sohan, Alastair Tse, and Anil Madhavapeddy.
Technical report (CUED/F INFENG/TR.533) at Cambridge University Engineering Department.
Using camera-phones to interact with context-aware mobile services
Eleanor Toye Scott, Anil Madhavapeddy, Richard Sharp, Dave Scott, Alan Blackwell, and Eben Upton.
Technical report (UCAM-CL-TR-609) at University of Cambridge, Computer Laboratory.