Automatic camera control using unobtrusive vision and audio tracking

Abhishek Ranjan*, Jeremy Birnholtz, Ravin Balakrishnan, Dana Lee

*Corresponding author for this work

Research output: Chapter in Book/Report/Conference proceedingConference contribution

13 Scopus citations


While video can be useful for remotely attending and archiving meetings, the video itself is often dull and difficult to watch. One key reason for this is that, except in very high-end systems, little attention has been paid to the production quality of the video being captured. The video stream from a meeting often lacks detail and camera shots rarely change unless a person is tasked with operating the camera. This stands in stark contrast to live television, where a professional director creates engaging video by juggling multiple cameras to provide a variety of interesting views. In this paper, we applied lessons from television production to the problem of using automated camera control and selection to improve the production quality of meeting video. In an extensible and robust approach, our system uses off-the-shelf cameras and microphones to unobtrusively track the location and activity of meeting participants, control three cameras, and cut between these to create video with a variety of shots and views, in real-time. Evaluation by users and independent coders suggests promising initial results and directions for future work.

Original languageEnglish (US)
Title of host publicationProceedings - Graphics Interface 2010
Number of pages8
StatePublished - Aug 20 2010
Event36th Graphics Interface Conference, GI 2010 - Ottawa, ON, Canada
Duration: May 31 2010Jun 2 2010

Publication series

NameProceedings - Graphics Interface
ISSN (Print)0713-5424


Other36th Graphics Interface Conference, GI 2010
CityOttawa, ON


  • Automated camera control
  • Computer vision
  • Meeting capture
  • Video

ASJC Scopus subject areas

  • Computer Graphics and Computer-Aided Design

Fingerprint Dive into the research topics of 'Automatic camera control using unobtrusive vision and audio tracking'. Together they form a unique fingerprint.

Cite this