Globalisation and an increasingly multidisciplinary collaboration drive the need for efficient ways to collaborate within distant teams. Although existing computer-supported cooperative work (CSCW) systems, such as collaborative video-conferencing and shared desktops, are suitable for certain tasks, most of these systems have two major deficiencies. First they support the situation of face to face meetings poorly. With a few exceptions, they use a desktop metaphor and this metaphor loses many of the subtleties of interpersonal communication since this does not model or represent the spatial relationships of the participants. Second, again because of the imposition of a desktop metaphor, many systems deal badly with data sets and workspaces that are inherently 3D in nature. What is missing is the ability to naturally gesture towards and manipulate the objects of discussion, and to observe the body language of the collaborators in the spatial context that it is responsive to. Thus, we lose subtle gestures such as gaze being drawn to an object of common interest, or gaze being used by a speaker to indicate whom they want to speak next as they finish speaking. This article serves two purposes. Firstly, it outlines the requirements for a class of CSCW tools that would focus on supporting closely coupled collaborative activity around shared