Note: To play any of the
videos, just click on the associated thumbnail. I am in the process
of uploading full-resolution copies of these videos onto YouTube. If the
video that you want to play is still in low-res Flash, contact
me and I will attempt to upload it ASAP. For your information, all of the
videos that I have uploaded onto YouTube appear in the channel:
wasbuxton.
|
|
|
|
Door Mouse
(1 min. 10 sec.)
University of Toronto
1990 |
This is work done by my student
Andrea Leganchuk. Sadly, she never published it. "Doorstates"
and the doormouse are an early example of "ambient intelligence", or what
we call "proximal sensing," or "reactive environments." The problem
addressed is this, when you have the "infomration highway right to your
door, how do you control access and availablity?" Our general design
philosophy was, whenever posible, use the same social protocols for virtual
visitations as for physical ones. Hence, if my door was the prime
cue to communicate my accessibility to someone in the physical corridor,
then let that same physical door be the mechanism to communicate my accessibility
to those coming down the virtual corridor. Hence, Andrea's wonderful
solution was to instrument my physical door with the guts of a second
mouse that was connected to my computer, and then drove the resulting application,
"Doorstates."
Ref:
Buxton, W. (1997). Living in
Augmented Reality: Ubiquitous Media and Reactive Environments. In K. Finn,
A. Sellen & S. Wilber (Eds.). Video
Mediated Communication. Hillsdale, N.J.: Erlbaum, 363-384. An earlier
version of this chapter also appears in Proceedings of Imagina '95,
215-229. |
 |
Hydra
(4 min. 11 sec.)
University of Toronto
1992 |
This video demonstrates the use
of what could be called "video surrogates" to afford the use of spatial
cues that are found in face-to-face conversation, and which support gaze
awareness, turn taking, and parallel conversations.
Ref:
Sellen, A., Buxton, W. &
Arnott, J. (1992).Using
spatial cues to improve videoconferencing. Proceedings of CHI '92,
651-652. |
 |
"Extra Eyes"
(1 min. 53 sec.)
University of Toronto
1995 |
Most video conferencing systems focus on
the remote person or object that are central to the purpose of the meeting.
In comparison to same-place meetings, one has very little sense of the
background or periphery. This system shows one approach of maintaining
an awareness of both the foreground and the periphery.
Ref:
Yamaashi, K., Cooperstock,
J., Narine, T. & Buxton, W. (1996). Beating
the limitations of camera-monitor mediated telepresence with extra eyes
Proceeding
of CHI '96, 50-57 |
 |
ConEd: Video Conversation Analysis
Tool
(1 min. 24 sec.)
University of Toronto
1992 |
This is a brief demonstration
of some of the techniques that we developed as part of the Telepresence
and Cavecat projects to analyse data recorded on audio and video
tape. It is of additional interest in that it was one of the first
applications into which we incorporated radial or "marking" menus.
Ref:
Kurtenbach, G. & Buxton, W. (1994). User
learning and performance with marking menus. Proceedings of CHI
'94, 258-264. |
|