Archive for the ‘Tech’ Category

Posted: January 6, 2011 in Art, Tech

I love these guys, Im not sure how to describe them but they do like interactive, artsy, tech stuff. Such as the amazing GML or graffiti markup language stuff which I think I blogged once but might dig out again. One of the members is Theo Watson who has always been a bit of an inspiration as an interactive designer, he was the first person I ever came across who seemed to be using programming in a creative way, or rather, to achieve a product which the user in turn could be creative with.

Anyway, fat labs are pretty cool imo.

Dedicated to enriching the public domain one mutha-fuckin LOL at a time
The Free Art and Technology Lab is an organization dedicated to enriching the public domain through the research and development of creative technologies and media. The entire FAT network of artists, engineers, scientists, lawyers, musicians and Bornas are committed to supporting open values and the public domain through the use of emerging open licenses, support for open entrepreneurship and the admonishment of secrecy, copyright monopolies and patents. All F.A.T. shine to Jonah Peretti, Mike Frumin, and the original open source r&d bad boys, the Eyebeam OpenLab Itsyaboy!
“What is FAT? Hard to say but somehow I feel it lies somewhere between these two images:”

Such a brilliant and inspiring talk. Totally mind blowing, I think I have watched it 3 times now and im still complately amazed!
The future of the internet is very exciting to me, and so to listen to this Kevin Kelly (the dude who started The Well, the first interactive bulletin boards on the www) talk about his findings in analizing the data so far and then projecting what this would mean in the future, is a very special thing. Well worth a watch…

The Future of Interaction…

Posted: January 6, 2011 in Tech

I came across this article and decided to post the whole thing up here because its so awesome.

The Future of User Interaction

By Demetrius Madrigal and Bryan McClain

Published: January 5, 2011

“We must keep in mind the new technologies that are currently under development and could influence the fundamental ways in which people interact with the products we design and develop.”

We think it’s appropriate to kick off the new year with an examination of what the future holds for user experience and product development. To stay ahead of the curve when it comes to user research and UX design, we must keep in mind the new technologies that are currently under development and could influence the fundamental ways in which people interact with the products we design and develop. We’ve seen the advent of such disruptive technologies repeatedly throughout time, including automobiles, telecommunications—radio, telephones, and television—and the personal computer. In this column, we’ll describe several new technologies that have the potential to change how we interact with technology and the world. Some of these technologies may be many years away from maturity, but they are definitely going to have massive impact in years to come.

Brain-Computer Interface

“The Brain-Computer Interface (BCI) has been under development since the 1970s. Its goal is to enable people to provide input and control systems using thought alone.”

The Brain-Computer Interface (BCI) has been under development since the 1970s. Its goal is to enable people to provide input and control systems using thought alone. To accomplish this, researchers like Philip Kennedy and John Donaghue have been studying ways to measure brain activity and convert the data into input for computer systems. For example, over the years, different applications have had success with restoring vision to  blind patients, control of prosthetic limbs,control of  robotics, speech generation, gaming, and cognitive imaging, among others.

Recent developments have attempted to make BCIs more flexible and less invasive, progressing from neuronal implants in the brain to  headsets using EEG (electroencephalography) technology like that shown in Figure 1. There are still many hurdles to overcome before this technology can truly reach its full potential, but it’s easy to see how BCIs could completely transform the way we interact with machines.

Figure 1—EEG headset

EEG headset

Instead of providing input using error-prone touchscreens that force us to make awkward gestures with our hands or QWERTY keyboards that were deliberately designed to slow our rate of input, we could use our thoughts to directly provide input to computing systems. Rather than typing and using a mouse, which can result in our developing carpal tunnel syndrome or other repetitive stress injuries, we could experience the ease of inputting text and commands directly. We would no longer have to fumble with tiny mobile device user interfaces that offer limited interaction models.

Instead, we would be able to do just about anything we could imagine from wherever we happen to be, through completely new modes of interaction. We could enjoy complete gaming experiences on our mobile phones and notebook computers, without any need for special controllers. We’d be able to input destinations into our cars’ navigation systems or select music while driving without taking our hands from the steering wheel or our eyes from the road. We might even be able to communicate soundlessly and create digital art that is a direct expression of our imaginations.

Of course, with such dramatic change would also come a complete shift in thinking when it comes to product design. For example, it would be essential to address design challenges like accidental activation. How much of our input should be automatic and how much should be intentional—and how can we distinguish between the two? Given all the chatter in our heads, how can we organize our thoughts into coherent user interactions? When people are communicating through their thoughts, how can we ensure that their fleeting thoughts don’t become unintended messages? What are the safety implications of operating heavy machinery like driving a car?

Although BCIs can free up the hands, does that mean people would be doing nothing but thinking? If people were controlling games with their minds, what would they be doing with their hands? Should their hands be doing nothing, or should they use them in some way to enhance the feeling of realism—like pretending to hold a gun or a guitar? How many things can a person do simultaneously? If a person is playing a complicated piano piece, could he also simultaneously edit the sheet music as he plays?

The possibilities for the applications of a Brain-Computer Interface are nearly endless, which could open up a whole new world of interaction design. It will be exciting to see what new interaction models emerge as this new technology matures.

Bionic Vision

“A bionic contact lens … would give us an entirely new way of interacting with technology.”

Typically, when people refer to wearable displays, they’re thinking of awkward glasses, goggles, or some other head-mounted display that provides a user with some kind of heads-up display. However, engineer Babak Parviz and his colleagues have envisioned something completely different. For several years, they have been working on a bionic contact lens, shown in Figure 2, that would enable the kind of augmented vision we’ve seen in science-fiction fantasies for decades. If Parviz succeeds, he would give us an entirely new way of interacting with technology. No longer would we be tied to our desks or need to lug around notebook computers. Nor would we be limited to using the small displays of portable devices like smartphones or tablets. This technology would free us from all such constraints and allow us to consume information in an entirely different way.

Figure 2—A bionic contact lens

Bionic contact lens

With these future possibilities would come new user experience challenges. We must discover how this new display medium would alter the way in which people interact with technology. We might be, and likely would be, surprised to discover what people would want to do with bionic vision and what uses they would avoid. It seems obvious that people would want to quickly and easily access information about the world around them—for example, looking up bus and train schedules as they hustle down busy city streets. But we must learn things like whether people would want to watch movies in full IMAX resolution while sitting on a plane, traveling over long distances.

We must also discover the most effective way of presenting information that is easily consumable while people are interacting with the world. While having instant information available at all times sounds exciting, it’s entirely possible users might absolutely hate viewing an augmented-reality information display incessantly, while trying to live their lives.

There would be challenges to overcome. Would people cling to traditional displays for whatever reason? Would people experience motion sickness from having information appear on a static display that doesn’t move with their environment as they turn their heads? How can we avoid safety issues that might arise from people being distracted from the real world as they walk down a street or drive their car? If people were viewing displays that only they could see, how would they be able to share what they’re seeing with other people? Answering such questions is essential to determining how people might actually put this new technology to use.

Affective Computing

Affective computing is an effort to instill machines with the ability to recognize and respond to emotions, simulating empathy.”

Historically, people have perceived computer systems as aloof, emotionless, and insensitive entities—like HAL in the visionary 2001: A Space Odyssey. While researchers and engineers like John Donaghue, Philip Kennedy, and Babak Parviz are seeking to change the way we interact with machines, Rosalind Picard’s goal is to change the way machines interact with us, through the development of  affective computing. Affective computing is an effort to instill machines with the ability to recognize and respond to emotions, simulating empathy. Rosalind Picard heads up the Affective Computing Research Group at the prestigious MIT Media Lab. As a discipline, affective computing relies upon groundbreaking research such as Paul Ekman’s work in involuntary microexpressions, Klaus Scherer’s work in vocal emotional expression, and nearly a decade of research in  galvanic skin response, which is an extremely reliable method of measuring physiological excitation and, thus, an excellent means of determining the intensity of emotions users experience.

An affective computing system that combined these different means of affective measurement could respond to a user’s current emotional state—for example, safety systems could respond to fear. Other planned applications of affective computing include systems that display emotion, resulting in more realistic digital pets or representatives in virtual systems for customer service and technical support.

Affective computing could significantly alter our relationship with technology. What we currently treat as a tool, we may someday interact with like a friend. Different brand identities might have a different feel. At some point, interaction design might encompass the design of systems that have different personality types and responses that range from stoic to empathetic to humorous. This would open the door to a new frontier of hedonic design. A system that could sense the emotional state of a user might learn to recognize positive emotional states in response to certain content or methods of display. The system could then adjust its output accordingly, to maximize the enjoyment of the user. Gaming systems incorporating affective computing could adjust their difficulty in response to a user’s frustration. Music players could create playlists that conform to a user’s mood.

The implications of affective computing for user experience research are quite significant. By providing us with a means of reliably measuring a user’s emotional state, we could identify which product or user interface designs result in frustration or delight, allowing us to better support the design of a positive user experience.

As such affective computing capabilities develop, we must identify important user considerations. How much emotional response and empathy would be too much? Would users be comfortable with emotional responses from computers? How variable would different users’ emotional responses be to the same user experience? Are there certain personality types that are more popular than others? Do mood changes alter users’ preferences for interactions? Are emotional responses to design elements consistent, or do they change over time? How can software systems alter their design on the fly to maximize users’ enjoyment? Successful implementation of affective computing would require an understanding of all of these issues.


“We have yet to see what form these technologies might take or how their developers might bring them to market. We don’t yet know what such products’ capabilities might ultimately be.”

In this column, we’ve taken a peek at the future. We’ve discussed possibilities for the future, but we have yet to see what form these technologies might take or how their developers might bring them to market. We don’t yet know what such products’ capabilities might ultimately be. Also, we haven’t yet touched on the ethical and social implications of more intimately connecting man and machine.

Transformative technologies such as these may have far-reaching impacts on how we interact with each other and with machines, how we engage in our professions, as well as how we preserve our health and safety. Our global society must determine how to best implement these technologies in useful and responsible ways.

Prospects for these new technologies are very exciting, but also a little daunting—and even more so when we start to imagine a world in which all of them exist at once. Imagine riding down a mountain on a bike, with your contact-lens display showing you three different route options, affective computing algorithms recommending which one you would enjoy most, and selecting your route with your mind. Such scenarios speak to a future in product design and user experience that would be both challenging and thrilling. We’re excited to see what types of user experiences people’s imaginations might conceive of once these technologies remove some of the limitations that bound user experience today. Let us know what you would design with these capabilities. What can you imagine?

KORG Electribe for the iPad

Posted: January 6, 2011 in Music, Tech

Korg’s ELECTRIBE·R has been coveted by dance musicians for ten years running. Today, this indispensable instrument has been reborn as a dedicated iPad app – the KORG iELECTRIBE virtual analog beatbox!
With the iELECTRIBE, sound creation is easy, fun and intuitive – just as on a vintage analog synthesizer. Simply choose a part (or sound) and touch the 16-step sequencer to quickly build a groove. This easy-to-use interface combines forces with the iPad’s 9.7” multi touch display to offer an intuitive “hardware” feel, sure to inspire and captivate the imagination of any user.The iELECTRIBE faithfully recreates the historic ELECTRIBE·R’s entire sound engine and sequencer capabilities. In addition, it provides advanced Motion Sequencing, eight supercharged effect types, plus 64 new preset patterns that instantly allow the user to create music in a wide variety of dance and electronic music styles. Get started today!

iPad Screenshot 2

Cute iPhone game with litle point! 🙂

posted May 13th, 2010 8:37 PM UTC by Eli Hodapp in $2.99, Games, iPhone games, iPod touch games

Cosmovox vs Electribe MX

Posted: January 6, 2011 in Music, Tech

Self Folding Paper

Posted: January 6, 2011 in Art, Tech

A self-folding sheet becomes an origami boat and an origami plane. Credit: The Harvard Microrobotics Lab