Escolar Documentos
Profissional Documentos
Cultura Documentos
0703-o3006
1541-4922 2007 IEEE
Published by the IEEE Computer Society
Applications
MVV technology could lead to exciting new applications in areas such as education, medicine,
surveillance, communication, and entertainment. It could also lead to a mass-media shake-up and the
birth of a new industry, especially in the mobile domain. Furthermore, researchers will also need to
examine surround sound with a fresh perspective to accompany the video style. MVV can also
profoundly affect telecommunication, given that telecommunications ultimate goal is highly effective
interpersonal information exchange.
For instance, media sports coverage technology keeps evolving. In the past, only a few TV channels
aired the games that interested people. Now audio and video coverage can be delivered over the
Internet or broadcast in HDTV format. Technology has always dazzled sports fans. Instant replays,
introduced in the early 1960s, added a new dimension that in-stadium fans couldnt see, and
miniature cameras let viewers see what referees see on the field. As MVV technology matures, we can
expect a revolution in coverage of sportscar racing, soccer, football, basketball, and so on. With
multiple cameras capturing and broadcasting the scene live to viewers and letting them rotate the
viewing angle, sports viewing could become a whole new concept.
Current videoconferencing systems provide a fixed view of the remote scene, so they dont give you
the feeling of being there. Multi-view video could have a broad impact on such systems. One
important feature of future communications will be interactivity with stereoscopic and 3D vision, which
make you feel more as if youre present in the scene. In a videoconferencing scenario, participants at
different geographical sites could meet virtually and see one another in free viewpoint video or 3DTV
style.
Surveillance and remote monitoring of important sites, such as critical infrastructures, traffic, parking
lots, and banks, could also benefit from this technology because it can provide coverage of very large
areas from multiple angles. Other potential application areas include entertainment (such as concerts,
multiuser games, and movies), education (such as digital libraries and archives, training and
instruction manuals with real video, and surgeon training), culture (such as zoos, aquariums, and
museums), and archiving (such as scientific archives, national treasures, and traditional
entertainment).
IEEE Distributed Systems Online (vol. 8, no. 3), art. no. 0703-o3006
Research issues
An MVV system consists of components for data acquisition, compression, and delivery. The
acquisition component captures videos from multiple cameras and obtains the acquisitions
parameters. The processing part analyzes the acquired data, extracts features of it, and compresses it
for delivery and storage. On the receiving side, decoding and display devices reconstruct the view in
either two or three dimensions, depending on the devices capabilities.
IEEE Distributed Systems Online (vol. 8, no. 3), art. no. 0703-o3006
IEEE Distributed Systems Online (vol. 8, no. 3), art. no. 0703-o3006
Projects
Because multi-view video is a new and widely applicable research area with a broad range of open
problems, numerous related research efforts are under way worldwide. In Europe, the Digital
Stereoscopic Imaging and Application (DISTIMA) project addressed the production, presentation,
coding, and transmission of digital stereoscopic video signals over integrated broadband
communications networks. Another European research project, the Package for New Operational
Autostereoscopic Multiview System ( PANORAMA), has aimed to facilitate the hardware and software
development of an MVV autostereoscopic telecommunication system. The Advanced Threedimensional Television System Technology (ATTEST) project aims to design an entire 3D-video chain,
including content creation, coding, transmission, and display. Mitsubishi Electric Research
Laboratories, Carnegie Mellon Universitys computer vision lab, Kyoto University, Heinrich Hertz
Institute in Germany, and the blue-c project are pursuing similar endeavors.
IEEE Distributed Systems Online (vol. 8, no. 3), art. no. 0703-o3006
Outlook
A workgroup of the International Organization for Standardizations Motion Picture Expert Group has
been exploring 3D audiovisual technology. The 3DAV has discussed various applications and
technologies in relation to the term multi-view video. A multi-view profile is available in the MPEG-2
standard, which was defined in 1996 as an amendment for stereoscopic TV. The MVP extends the wellknown hybrid coding toward exploitation of inter-view/channel redundancies by implicitly defining
disparity-compensated prediction; however, it doesnt support interactivity. MPEG-4 version 2 includes
the Multiple Auxiliary Component, defined in 2001. MACs basic idea is that grayscale shape is used
not only to describe the video objects transparency but also can be defined in a more general way.
MACs are defined for a video object plane on a pixel-by-pixel basis and contain data related to the
video object, such as disparity, depth, and additional texture. Since 2003, MPEG has also accelerated
its work on MVV coding standards. The Multiview Video Coding initiative has passed MPEGs call-forproposals stage. The proposals were based on the H.264/AVC video coding standard. Thus, the MVC is
currently being developed and standardized as an extension of this standard in a joint ad hoc group on
MVC (AHG on MVC) in JVT. For more information on the MPEG standardization efforts, see
www.chiariglione.org/mpeg/working_documents.htm.
MVV-based products are expected to appear in two to three years. Watching television passively might
soon be a thing of the past.
References
1.
2.
3.
4.
5.
6.
7.
8.
9.
10.
W. Matusik and H. Pfister, 3D TV: A Scalable System for Real-Time Acquisition, Transmission,
and Autostereoscopic Display of Dynamic Scenes, ACM Trans. Graphics, vol. 23, no. 3, 2004,
pp. 814824.
T. Kanade, H. Saito, and S. Vedula, The 3D Room: Digitizing Time-Varying 3D Events by
Synchronized Multiple Video Streams, tech. report CMU-RI-TR-98-34, Carnegie Mellon Univ.,
1998.
T. Kanade, P.W. Rander, and P.J. Narayan, Virtual Reality: Constructing Virtual Worlds from
Real Scenes, IEEE Multimedia, vol. 4, no. 1, 1997, pp. 3447.
M. Gross et al., blue-c: A Spatially Immersive Display and 3D Video Portal for Telepresence,
Proc. ACM Intl Conf. Computer Graphics and Interactive Techniques (SIGGRAPH 03), ACM
Press, 2003, pp. 819827.
S. Moezzi et al., Immersive Video, Proc. Virtual Reality Ann. Intl Symp. (VRAIS 96), IEEE CS
Press, 1996, pp. 1724.
H. Saito, S. Baba, and T. Kanade, Appearance-Based Virtual View Generation from
Multicamera Videos Captured in the 3-D Room, IEEE Trans. Multimedia, vol. 5, no. 3, 2003,
pp. 303316.
A. Smolic and D. McCutchen, 3DAV Exploration of Video-Based Rendering Technology in
MPEG, IEEE Trans. Circuits and Systems for Video Technology, vol. 14, no. 3, 2004, pp. 348
356.
H. Cheng, Temporal Registration of Video Sequences, Proc. 2003 IEEE Intl Conf. Acoustics,
Speech, and Signal Processing (ICASSP 03), vol. 3, IEEE Press, 2003, pp. III 48992.
K. Mueller et al., Coding of 3D Meshes and Video Textures for 3D Video Objects, Proc.
Picture Coding Symp. (PCS 04), 2004.
O. Javed and M. Shah, Tracking and Object Classification for Automated Surveillance, Proc.
7th European Conf. Computer Vision (ECCV 02), Springer, 2002, pp. 343357.
Ishfaq Ahmad is a professor of computer science and engineering at the University of Texas at
Arlington. Contact him at iahmad@cse.uta.edu.
IEEE Distributed Systems Online (vol. 8, no. 3), art. no. 0703-o3006
Related Links
x
x
x
IEEE Distributed Systems Online (vol. 8, no. 3), art. no. 0703-o3006