- The 2014 World Cup kicked off yesterday with a futuristic twist on the opening ceremonies. A paraplegic kicked a soccer ball using an exoskeleton designed by the Walk Again Project:
The exoskeleton — a system comprising a helmet implanted with a microchip that sticks out from the underside; a T-shirt loaded with sensors; metal leg braces; and a battery worn in a backpack — is set in motion when the user envisions himself making the kick. The chip translates those electronic commands to a digital language that powers the skeleton, which then moves accordingly. The T-shirt vibrates to enhance the user’s sensation of movement (and eliminate the need to look at his feet to see if he’s stepping forward).
- Unfortunately, as io9 reports, the moment was not well-covered by TV networks:
Talk about dropping the ball. Earlier today, Juliano Pinto — a 29 year-old paraplegic — successfully kicked off the 2014 FIFA World Cup by using a mind-controlled exoskeleton. But sadly, most TV networks failed to show it.
After months of hype, the official broadcast of the opening ceremonies showed only a fraction of it, while some TV networks missed the event altogether. Commentators criticized the organizers for casting aside the moment in favor of performing acts.
- Thomas Frey at the Futurist Speaker blog forecasts the coming AI crash wars:
The invasion of high-frequency trading machines is now forcing capitalism far away from anything either Adam Smith or the founders of the NYSE could possibly find virtuous.
We’re not about to let robots compete in the Olympics, driverless cars race in the Indianapolis 500, or automated machines play sports like football, basketball, or baseball. So why is it we allow them to play a role in the most valuable contest of all, the world wide stock exchange?
With crude forms of AI now entering the quant manipulator’s toolbox, we are now teetering dangerously close to a total collapse of the stock market, one that will leave many corporations and individuals financially destitute.
- Microsoft has announced their version of apple’s Siri virtual assistant. Named Cortana, after the AI character from the Halo video game series, she is coming to Windows smartphones, and as Brad Molen at engadget reports, developers programmed her with a distinct personality:
Confident, caring, competent, loyal; helpful, but not bossy: These are just some of the words Susan Hendrich, the project manager in charge of overseeing Cortana’s personality, used to describe the program’s most significant character traits. “She’s eager to learn and can be downright funny, peppering her answers with banter or a comeback,” Hendrich said. “She seeks familiarity, but her job is to be a personal assistant.” With that kind of list, it sure sounds like Hendrich’s describing a human. Which is precisely what she and her team set out to do during Cortana’s development; create an AI with human-like qualities.
Microsoft’s decision to infuse Cortana with a personality stemmed from one end goal: user attachment. “We did some research and found that people are more likely to interact with [AI] when it feels more human,” said Hendrich. To illustrate that desired human-machine dynamic, Hendrich pointed to her grandmother’s experience with a Roomba vacuum: “She gave a name and a personality to an inanimate object, and it brought her joy.” That sense of familiarity is exactly what Microsoft wants Window Phone users to feel when interacting with Cortana on their own devices.
- In case you haven’t already heard, scientists have implanted false memories into the brains of mice.
Scientists have created a false memory in mice by manipulating neurons that bear the memory of a place. The work further demonstrates just how unreliable memory can be. It also lays new ground for understanding the cell behavior and circuitry that controls memory, and could one day help researchers discover new ways to treat mental illnesses influenced by memory.
- The inevitable Total Recall references have already appeared. Others have gone with Inception as the pop culture touchstone.
- I recently discovered the Augmented Reality Trends website. Some noteworthy posts: How augmented reality aids advertising.
Augmented reality blurs the line between the virtual and real-world environment. This capability of augmented reality often confuses users, making them unable to determine the difference between the real world experience and the computer generated experience. It creates an interactive world in real-time and using this technology, businesses can give customers the opportunity to feel their products and service as if it is real right from their current dwelling.
AR technology imposes on the real world view with the help of computer-generated sensory, changing what we see. It can use any kind of object to alter our senses. The enhancements usually include sound, video, graphics and GPS data. And its potentials are tremendous as developers have just started exploring the world of augmented reality. However, you must not confuse between virtual reality and augmented reality, as there is a stark difference between them. Virtual reality, as the name suggests, is not real. It is just a made up world. On the other hand, augmented reality is enhancing the real world, providing an augmented view of the reality. The enhancements can be minor or major, but AR technology only changes how the real world around the user looks like.
And a profile of SeeMore Interactive and their work on augmented reality shopping:
Augmentedrealitytrends.com: Why augmented reality and why your prime focus is on retail industry?
SeeMore Interactive: We recognize the importance of merging brick-and-mortar retail with cloud-based technology to create the ultimate dynamic shopping experience. It’s simply a matter of tailoring a consumer’s shopping experience based on how he or she wants to shop; the ability to research reviews, compare prices, receive new merchandise recommendations, share photos and make purchases while shopping in-store or from the comfort of their home.
- Brian Matchick at Geek Exchange writes about how Deep Learning brings A.I. one step closer to Hal, Skynet, and the Matrix:
Deep learning is based on neural networks, simplified models of the way clusters of neurons act within the brain that were first proposed in the 1950s. The difference now is that new programming techniques combined with the incredible computing power we have today are allowing these neural networks to learn on their own, just as humans do. The computer is given a huge pile of data and asked to sort the information into categories on its own, with no specific instruction. This is in contrast to previous systems that had to be programmed by hand. By learning incrementally, the machine can grasp the low-level stuff before the high-level stuff. For example, sorting through 10,000 handwritten letters and grouping them into like categories, the machine can then move on to entire words, sentences, signage, etc. This is called “unsupervised learning,” and deep learning systems are very good at it.
- This Economist article looks at predictive policing and American company PredPol (amazingly, a sly sub-section heading is the only reference to the book or film Minority Report and it’s pre-crime unit and pre-cog mutants.):
Intelligent policing can convert these modest gains into significant reductions in crime. Cops working with predictive systems respond to call-outs as usual, but when they are free they return to the spots which the computer suggests. Officers may talk to locals or report problems, like broken lights or unsecured properties, that could encourage crime. Within six months of introducing predictive techniques in the Foothill area of Los Angeles, in late 2011, property crimes had fallen 12% compared with the previous year; in neighbouring districts they rose 0.5% (see chart). Police in Trafford, a suburb of Manchester in north-west England, say relatively simple and sometimes cost-free techniques, including routing police driving instructors through high-risk areas, helped them cut burglaries 26.6% in the year to May 2011, compared with a decline of 9.8% in the rest of the city.
- The BBC web site has published an article on the cities of the future:
Although they may all look very different, the cities of the future share a new way of doing things, from sustainable buildings to walkable streets to energy-efficient infrastructure. While some are not yet complete – or even built – these five locations showcase the cutting edge of urban planning, both in developing new parts of an existing metropolitan area and building entirely new towns. By 2050, it is forecast that 70% of the world’s population will live in cities. These endeavours may help determine the way we will live then, and in decades beyond.
- This piece from Vice’s Motherboard examines Bill Gates’ nuclear power company TerraPower and alternative nuclear fuel thorium:
Mention thorium—an alternative fuel for nuclear power—to the right crowd, and faces will alight with the same look of spirited devotion you might see in, say, Twin Peaks and Chicago Cubs fans. People love thorium against the odds. And now Bill Gates has given them a new reason to keep rooting for the underdog element.
TerraPower, the Gates-chaired nuclear power company, has garnered attention for pursuing traveling wave reactor tech, which runs entirely on spent uranium and would rarely need to be refueled. But the concern just quietly announced that it’s going to start seriously exploring thorium power, too.
- Unsurprisingly, a porno movie filmed using Google Glass has already wrapped:
Google might have put the kibosh on allowing x-rated apps onto Glass (for now) but that hasn’t stopped the porn industry from doing what they do best: using new technology to enhance the, um, adult experience. The not yet titled film stars James Deen and Andy San Dimas.
- Speaking of Google, the company’s research department recently announced details about their Machine Vision visual recognition program:
There has always been a basic split in machine vision work. The engineering approach tries to solve the problem by treating it as a signal detection task using standard engineering techniques. The more “soft” approach has been to try to build systems that are more like the way humans do things. Recently it has been this human approach that seems to have been on top, with DNNs managing to learn to recognize important features in sample videos. This is very impressive and very important, but as is often the case the engineering approach also has a trick or two up its sleeve.
- From Google Research:
We demonstrate the advantages of our approach by scaling object detection from the current state of the art involving several hundred or at most a few thousand of object categories to 100,000 categories requiring what would amount to more than a million convolutions. Moreover, our demonstration was carried out on a single commodity computer requiring only a few seconds for each image. The basic technology is used in several pieces of Google infrastructure and can be applied to problems outside of computer vision such as auditory signal processing.