The Earth Theater:
The Earth Theater at the Carnegie Museum of Natural History is a Panoromic SkyVision Theater, a product of SkySkan, Inc. The theater has a 40 foot diameter, a 210 degree horizontal viewing angle, a 30 degree vertical viewing angle, and seats an audience of 68. The theater's five digital light projectors, slide projectors, lighting instuments, and 4.1 channel surround sound are driven by SkySkan's Spice Multimedia Control System. The SkyVision system was specifically designed for the playback of pre-rendered full-dome & panoromic digital video.

ETC System Diagram
Click to view
ETC Panoromic Real-time Interaction System:
The Entertainment Technology Center's system consists of two off-the-shelf 866Mhz Intel Pentium III Computers: one for serial communication with Spice and rendering graphics to all five projectors, and a second for sound playback and interactive inputs. The rendering computer employs five ATI Radeon video cards on the PCI bus. The sound/interaction computer uses a Creative Labs Sound Blaster Live! card in order to create three-dimensional spatialized audio. Audience interaction is done through an active infra-red camera, coupled with an ImageNation PXC200 frame grabber, and an omni-directional microphone plugged into the SoundBlaster card.

The main software used on the project is Alice, a 3D authoring environment developed by the Stage3 Research Group at Carnegie Mellon University. Alice is designed for the rapid prototyping and development of interactive 3D graphics. Alice was intended to be used for material created for desktop computer and Head-Mounted display use. The ETC Team and Stage3 collaborated to adapt Alice for playback across multiple displays by having each camera object in the virtual environment output to a separate video card.

In order to create a seamless view across the five projected images, the ETC Team employed an alpha-channel gradient within the virtual world. This created issues where one camera could see another camera's blending object, so the visibility of the blending object was tied to the refresh rate of the projectors.

Since such an immense computational load was placed on a single computer, audio playback and processing of audio and video interations were outsourced to a separate machine. The remote sound and vision server programs were written by the students on the project.