I’ve been looking for a good answer for lifelogging for awhile, and have been anticipating the point at which the technology drops in price enough to make this possible for the average consumer. Google Glasses is the high-end answer, and this may just be the sort of thing that emerges at the low end.
The Memoto camera is a tiny camera and GPS that you clip on and wear. It’s an entirely new kind of digital camera with no controls. Instead, it automatically takes photos as you go. The Memoto app then seamlessly and effortlessly organizes them for you.
Incredible augmented reality technology that has real-world, everyday use: use cameras and computers to make seen what can’t been seen by the normal eye. In this case, it’s welding that benefits, but there is an obvious extension into many other fields.
As computers become even more powerful, it will be possible to do similar real-time HDR image processing with little more than a smartphone. There will be a point down the road where humans will be able to manufacture cheap and lightweight glasses that are capable of providing better imaging than our own eyes can manage. This will be one of several points where true augmentation of human capabilities begins — the very definition of cyborg. It’s not science-fiction, it’s the future.
Google’s Project Glass product lead Steve Lee walks us through his experience with the development of the company’s sci-fi-inspired eyewear–from his team’s “hundreds of variations and dozens of early prototypes” to his vision of the future.
By “wearable computing” I mean mobile computing where both computer-generated graphics and the real world are seamlessly overlaid in your view; there is no separate display that you hold in your hands think Terminator vision. The underlying trend as we’ve gone from desktops through laptops and notebooks to tablets is one of having computing available in more places, more of the time. The logical endpoint is computing everywhere, all the time – that is, wearable computing – and I have no doubt that 20 years from now that will be standard, probably through glasses or contacts, but for all I know through some kind of more direct neural connection. And I’m pretty confident that platform shift will happen a lot sooner than 20 years – almost certainly within 10, but quite likely as little as 3-5, because the key areas – input, processing/power/size, and output – that need to evolve to enable wearable computing are shaping up nicely, although there’s a lot still to be figured out.
And yet more from the wearable display front…the way that modern gadget production works seems to be very much a sort of punctuated equilibrium. There are ideas, then a breakthrough in sensor production and pricing causes an explosion of forms, which then settle down into a few winners (see: eReaders in 2010, Tablets in 2011, etc). I am very curious where this particular explosion is going to take us.
People who constantly reach into a pocket to check a smartphone for bits of information will soon have another option: a pair of Google-made glasses that will be able to stream information to the wearer’s eyeballs in real time.
According to several Google employees familiar with the project who asked not to be named, the glasses will go on sale to the public by the end of the year.