I was thinking “Minority Report” when I saw this in a science museum/tech expo around the December holidays. Although there didn’t seem to be a lot of functionality included within the demonstration panel the possibilities were endless. It seemed like this piece of technology could indeed revolutionise the way we use applications and technology.
Ok below here is mostly copied from the website, I am not going to write a new overiew on this technology when the guy who made it alreday explain it so well. Read on and check the website.
Cheers Willy for the early morning link!
“do do doooooooooo!”
Let’s start from the start.
“We introduce a simple technique that enables robust multi-touch sensing at a minimum of engineering effort and expense. It relies on frustrated total internal reflection (FTIR), a technique well known in the biometrics community for fingerprint image acquisition. It acquires true touch information at high spatial and temporal resolutions, and is scalable to very large installations.”
Excerpt from website.
Bi-manual, multi-point, and multi-user input on a graphical interaction surface.
While touch sensing is commonplace for single points of contact, multi-touch sensing enables a user to interact with a system with more than one finger at a time, as in chording and bi-manual operations. Such sensing devices are inherently also able to accommodate multiple users simultaneously, which is especially useful for larger interaction scenarios such as interactive walls and tabletops.
Since developing the FTIR (frustrated total internal reflection) technique, we’ve been experimenting with a wide variety of application scenarios and interaction modalities that utilize multi-touch input information. These are far more interesting than the typical poking-of-the-touchscreen or the gross silhouette gesturing found in most interactive installations of this scale. It is a rich area for research, and we are extremely excited by its potential.
The sensing technology is force-sensing, and provides unprecedented resolution and scalability, allowing us to create sophisticated multi-point widgets for applications large enough to accomodate both hands and multiple users.
The drafting table style implementation shown here measures 36″x27″, is rear-projected, and has a sensing resolution of ~0.1″ at 50Hz. Applications receive events and stroke information using the lightweight OSC protocol over UDP.
We’ll be working with other, more interesting form factors (both larger and smaller). Wouldn’t it also be nice to identify which finger is associated with each contact?