Editor’s note: This 2012 IBM 5-in-5 article is by Robyn Schwartz, associate
director of IBM Research Retail Analytics, and Dhandapani Shanmugam, solutions architect and Siddique A. Mohammed, software
architect of IBM Software Group Industry Solutions.
Within the next five years, your mobile device will let you touch
what you’re shopping for online. It will distinguish fabrics, textures, and
weaves so that you can feel a sweater, jacket, or upholstery – right through
the screen.
Shopping
moves to mobile
In
the U.S., online sales for Thanksgiving and Black Friday were up almost 20%
from the same two days in 2011, according to data from IBM Smarter Commerce.
Haptic devices such as gloves or “rumble packs” used in
gaming have existed for years. But we use them in closed environments where the
touch doesn’t actually connect to where we are in reality. We at IBM Research
think that in the next five years that our mobile devices will bring together
virtual and real world experiences to not just shop, but feel the surface of
produce, and get feedback on data such as freshness or quality.
It’s already possible to recreate a sense of texture through
vibration. But those vibrations haven’t been translated into a lexicon, or
dictionary of textures that match the physical experience. By matching
variable-frequency patterns of vibration to physical objects so, that when a
shopper touches what the webpage says is a silk shirt, the screen will emit
vibrations that match what our skin mentally translates to the feel of silk.
Vibrating air to feel like something solid
Using digital image processing and digital image
correlation, we can capture texture qualities in a Product Information
Management (PIM) system to act as that dictionary. Retailers could then use it
to match textures with their products and their products’ data – sizes,
ingredients, dimensions, cost, and any other information the customer might
expect. The dictionary of texture will also grow and evolve as we grow our
appetite, usage and understanding of this kind of technology.
We’re not ready to virtually high-five Tupac Shakur’s
hologram through a phone – yet. Soon though, the phone will be able to emit a
field of vibrations. Just millimeters from the screen. And the vibrations will
be subtle. Your phone won’t shake out of your hand, but will deliver a recognizable
sensation. Imagine shopping for a wedding dress on a
phone or tablet, and being able to feel the satin gown, or even the intricate
beading and buttons, or the lace on the veil.
Beyond the clothing rack
Starting in retail makes sense because we all intrinsically
understand the browsing and shopping experience. We all naturally respond to
and understand texture, from a soft pair of socks to a ripe piece of fruit.
The touch of somethingtranslated,
based on accumulated data in a database down to an end user’s mobile device
could also have the power to help us gain new understandings of our
environment. Take farming, for example. Farmers could use a mobile device
to determine the health of their crop by comparing what they’re growing to a
dictionary of healthy options that they feels through a tablet.
The technology could evolve beyond communicating textures retrieved
from some database, and toward real time touch translation gained from
accumulated user interaction with the technology. What is one of the first
things a doctor does when treating an injured patient? Touch the injury. The
patient could send a photo of an injury to let the doctor feel the injury
remotely to help make a faster diagnosis – before or perhaps instead of
visiting the doctor in person.
Five years may not seem like enough time to begin feeling
what you shop for through your smart phone. But look at how far the smart phone
has come. You can see the people you talk to via a camera. You can remotely adjust
your home’s thermostat, or set your home alarm system; pay your babysitter;
find your way to a local pizza restaurant; and watch a movie. Why can’t it also
help you stay in “touch” with your environment?
If you think cognitive systems will most-likely have the ability to touch, before augmenting the other senses, vote for it, here.
IBM thinks these cognitive systems will connect to all of
our other senses. You can read more about sight, smell, hearing, and taste technology
in this year’s IBM 5 in 5.