Multimedia Processing
and Communications Systems

Computing Japan visits the Multimedia Systems Laboratories of Fujitsu Laboratories, Ltd.

For this month's R&D Focus, Computing Japan visited Fujitsu's Multimedia Systems Laboratories in Kawasaki. Senior Editor Steven Myers reports on selected research projects of the MSL's Media Processing Lab and NetMedia Lab.

by Steven Myers

The multimedia systems laboratories, located near the head office of Fujitsu Laboratories, Ltd., in Kawasaki, is one of five major research centers that make up Fujitsu Laboratories, a wholly owned subsidiary of the famous parent company. Fujitsu Laboratories was created in 1962 by merging R&D sections that previously had been managed by separate technical divisions within Fujitsu. Today, the labs employ almost 1,600 persons and conduct research and development in fields ranging from high-end multimedia systems to low-level electron devices and materials.

The Multimedia Systems Laboratories, three other labs in Japan, and one lab in California round out Fujitsu Laboratories. The others are the Personal Systems Labs (Akashi), Electronic Systems Labs (Atsugi), Electron Devices and Materials Labs (Atsugi), and Fujitsu Laboratories of America, Inc. (Santa Clara, California). In addition to these five laboratories, Fujitsu also manages the Institute for Social Information Science (Numazu), which conducts research in areas such as biological information science and computer mathematics.

The Multimedia Systems Laboratories

Computing Japan's visit to the Fujitsu Kawasaki facilities was arranged by Mike Beirne, who is in the public relations department of Fujitsu Ltd. at the company's headquarters in Chiyoda-ku, Tokyo. Our thorough tour of the expansive Multimedia Systems Labs, which included numerous demonstrations, was capped off by a visit to the "museum" located in the head office building of Fujitsu Laboratories. The museum contains exhibits of many of the famous products and technologies developed by Fujitsu over the past fifty years.

The Multimedia Systems Laboratories' facilities are enormous; it was difficult to grasp the scope and scale of the work conducted there based on a single visit. The facilities are broken down into a dozen or so smaller labs, including the Media Integration Lab (natural language processing, optical character recognition, and multimedia databases), NetMedia Lab (software agents and Internet applications), Media Processing Lab (video servers and image processing), and System LSI Lab (LSI chips for image processing). All of the labs work to develop technologies in their respective areas, with an overall goal of producing systems for multimedia processing and communications.

In visiting research labs, it is not an uncommon situation that much is talked about, but little is shown. This was definitely not the case at the Multimedia Systems Labs. The ability of Fujitsu's researchers to explain their work and their willingness to demonstrate the systems they have built were impressive. At each lab, we were first given a thorough introduction to the research goals for the projects underway, then shown an impressive array of demonstrations of the experimental systems built by the project teams. In addition to supplying us with papers and technical reports, all of the researchers gave high-quality presentations and were quick to answer any questions about their work.

Described here are just three of the many projects we saw demonstrated at the Multimedia Systems Laboratories.

AIMS

Drawing upon its vast experience in developing technologies for VLSI chips, storage devices, ATM (asynchronous transfer mode) equipment, and software systems, a group at the Media Processing Lab is currently at work building a scaleable multimedia server. This server eventually will be used for advanced services such as VOD (video-on-demand), teleshopping, and remote learning. The immediate goal for this team is to develop an interactive visual information system that can distribute high-quality MPEG-2 videos to clients over an ATM network.

The AIMS (ATM-based Interactive Multimedia Server) system utilizes an architecture that provides a direct connection between the disks controlled by the server and the ATM interfaces. With this massively parallel architecture in place, 20 programs at speeds of 6MB per second each can be delivered through an ATM port. The system is fully scaleable; the size and speed can be increased or reduced simply by adding or removing of unit modules. (Each unit has the specifications shown in figure 1.) It is noteworthy that the system can handle up to 350 simultaneous streams per unit at a speed of 3.3MB per second.

The AIMS system uses an MPEG-2 decoder card implemented with a newly developed Fujitsu video decoder VLSI chip, the MB86365. This new card conforms to PCI bus specifications and is fully compatible with the MPEG-1 and -2 standards. (Figure 2 depicts a block diagram of the decoder.) Using a SPARClite RISC processor, the card provides flexible support for the MPEG system layer syntax. Although the AIMS server is currently being used only for experimental purposes, the team expects that the system will have a strong impact on broadcasting and communications in the near future.

ISHTAR

Another team at the Media Processing Lab is developing a computer vision system that can recognize the motion of objects. ISHTAR (Image Sequence Hardware for Temporal Analysis in Real-time) uses an optical flow algorithm for calculating the velocity of a moving object. Although the algorithm is relatively simple to implement, it has a prohibitively high computational cost on traditional systems. The ISHTAR team has therefore designed and built its own linear array processor, one that directly processes TV images in real time and thus allows for a more reasonable implementation of the optical flow algorithm.

The actual system consists of a video I/O board and several DSP (digital signal processing) boards. The video board digitizes NTSC analog signals from a video camera or VCR, separates the image and synchronous data, and transfers them to the DSP boards. Each DSP board contains two TMS320C40 DSP chips and can be connected in tandem with other DSP boards to build a linear array system.

One of the leaders of the ISHTAR project, Shigeru Sasaki, explained that in several experiments the system has proven capable of calculating velocities for image sequences in real time. In one of these experiments, the team used a version of ISHTAR consisting of a video board and four DSP boards to successfully do real-time calculations of the velocity of shifting and rotating objects. Sasaki went on to explain that Fujitsu is looking at applying the technology developed by the ISHTAR team to a variety of fields, most notably traffic surveillance.

DUET

The DUET team, headed by Ichiro Iida of the NetMedia Lab, aims to develop an agent-based personal communications system that integrates several diverse network technologies into a single architecture that is invisible to users. Under such an architecture, users would be able to employ any type of telephone, PDA (personal digital assistant), or PC to communicate with other persons on the network, regardless of the device types used by the receiver. The details of each physical network are encapsulated in a "service control network layer."

Cooperation among network nodes in the service control network is managed by a group of Personal Agents (PAs). Each user has his or her own PA, which maintains a set of information about that user, including location and preferred method of interface. To send information, a user need only specify the ID number for another user. The system then contacts the PA corresponding to the receiver, retrieves the information for the receiver, and converts the data to be sent into a suitable format.

Rather than having the user manually register his or her ID with the exact devices to be used, the goal is to have this task be performed dynamically by the network. One solution proposed by the group is to have each user carry a small wireless personal device that transmits the user's ID by radio signals. The network will then determine the person's location by polling the devices in each area, and automatically register the user to the appropriate devices.

Iida demonstrated some of the features of this architecture for us using a prototype system developed by the team that allows users to move around and perform different communication activities using PDAs and PCs. In one example, an Apple Newton with a wireless connection was used to send e-mail and make telephone calls to other users on the network. Iida showed us the information stored by the system for the users, and demonstrated how the information is updated automatically when a user changes his/her physical location. Iida notes that this type of architecture is especially suited for advanced communications networks that must integrate real-time voice and non-real-time messaging services.



Contact information
Fujitsu Laboratories Ltd.
Phone 044-777-1111
Fax 044-754-2580




Copyright 1996 Computing Japan