March 5, 2007 While some people just adore trying on shoes, for most of us, it’s a drag, but the application of new technologies by adidas looks set to free us from this tedium and save considerable time for everybody. This entertaining new form of shopping can be found on the Avenue des Champs Elysees, where adidas has opened its latest and most modern shop anywhere in the world. Customers can now try on a variety of models in front of a virtual mirror without changing their shoes. They can navigate through the collection by simply pointing at products on a computer screen.
Instead of trying on dozens of pairs, the customer simply stands in front of a virtual mirror. On his foot, he can see the shoe of choice (as long as it’s adidas) in the colour of choice. The virtual mirror was developed by researchers at the Heinrich-Hertz-Institut HHI in Berlin. Unlike a conventional mirror, it does not display a true reflection. Instead, a camera captures the customer’s feet and legs and displays them as a video scene on the monitor. The various shoe models are inserted into this picture. “Thanks to the 3-D image processing techniques developed at the HHI, the software is so fast that it can follow the customer’s movements in real time,” says Jurgen Rurainsky, one of the virtual mirror’s developers.
A second presentation area known as “Infospace”, also conceived by the HHI, presents shoes and apparel from the sportswear manufacturer in pictures, ad spots and brief movies. In contrast to the touch screen technique, the customer can navigate through the menu without touching anything at all. All he needs to do is point at the screen with his index finger from a distance of approximately 80 centimeters.
What makes it all possible is a “finger-tracking” system: A stereo camera on the ceiling photographs the finger and calculates its spatial position and the direction in which it is pointing. The information is passed on to a software program that moves and activates the objects on the monitor. If the customer wants to click on an object, for example a video scene, he simply keeps his finger pointing in the appropriate direction for a moment.
The challenge for the researchers was to program the system so skillfully that it would be fast enough to respond instantly to these motions. After all, it not only has to interpret the finger’s motion correctly within fractions of a second, but also translate the gestures into the appropriate commands without delay. “Another important goal was to make this technology understandable and easy to use,” explains Paul Chojecki, who tested the user-friendliness of the control-by-gesture system.
“But our goal for both presentation areas was to demonstrate that you can move around in the virtual world without recourse to technical aids such as data gloves or 3-D goggles.”
The researchers will be demonstrating how well their method works at the CeBIT fair in Hanover from March 15 to 21 (Hall 9, Stand B36).