Earlier this month the mobile-app game Pokémon Go was released in the U.S., and the game has been ubiquitous ever since. Aside from being a sudden pop culture phenomenon, the game’s success poses some significant implications. First of all, this is clearly a breakthrough moment for augmented reality. Pokémon Go is not the first augmented reality game, nor is it the most ambitious, but it has undoubtedly brought AR into mainstream consciousness. Secondly, the success of Pokémon Go has led me to reconsider all my previously held assumptions about the uses of mobile apps and gamification for interfacing with urban spaces. I have historically been cynical about the prospect of using mobile games or AR interfaces to interact with urban space, since they usually strike me as shallow and insignificant, typically resulting in a fleeting diversion like a flash mob dance party, rather than altering people’s perceptions of place in any lasting or meaningful way. Pokémon Go satisfies all the requirements of my earlier preconceptions, yet despite my best critical instincts, I really like the game.
The buzz about Pokémon Go had been building on various forums online, and after it was released it was virtually impossible to avoid Pokémon Go-related posts. Save for maybe 10 minutes with a friend’s Game Boy in the late 90s, I’ve never played a Pokémon game and I preemptively wrote off Pokémon Go as yet another cultural fad that I would never partake in or understand. Curiosity got the best of my wife, however, and she downloaded the app and we walked around our neighborhood to test it out. To my surprise, the game was a lot of fun; our familiar surroundings were now filled with digital surprises, and we were excited to see neighborhood landmarks and murals represented as Pokéstops, and wild Pokémon hanging out in the doorways of local shops. We meandered around discovering which of our local landmarks had been incorporated into the game, and each discovery increased my enjoyment of the app. Yes, the game is simple and shallow, but I was completely charmed. I downloaded the game so I could play, too.
Reactions to Pokémon Go have been as fascinating as the game’s widespread adoption. Many news articles sensationalized the inherent dangers of playing the game: distracted players wandering into traffic or off of cliffs, people’s homes being designated as Pokéstops and besieged by players, and traps being laid (using the games “lures”) to ambush and rob aspiring Pokétrainers. There have also been insightful critical analyses of the game. An early and oft-shared article by Omari Akil considered the implications of Pokémon Go in light of recent police shootings of black men, warning that “Pokemon Go is a death sentence if you are a black man“:
I spent less than 20 minutes outside. Five of those minutes were spent enjoying the game. One of those minutes I spent trying to look as pleasant and nonthreatening as possible as I walked past a somewhat visibly disturbed white woman on her way to the bus stop. I spent the other 14 minutes being distracted from the game by thoughts of the countless Black Men who have had the police called on them because they looked “suspicious” or wondering what a second amendment exercising individual might do if I walked past their window a 3rd or 4th time in search of a Jigglypuff.
Others questioned the distribution of Pokémon across neighborhoods, suggesting that poor or black neighborhoods had disproportionately fewer Pokémon and Pokéstops. Among urbanists, however, reaction to the game has been mixed. Mark Wilson at Fastcodesign declared that Pokémon Go “is quietly helping people fall in love with their cities“. Ross Brady of Architizer celebrated the game for sparking “a global wave of urban exploration“. Writing for de zeen, Alex Wiltshire boldly states that the game has “redrawn the map of what people find important about the world“. City Lab contributor Laura Bliss proclaimed “Pokémon Go has created a new kind of flaneur“.
Others have been more critical of the game, with Nicholas Korody at Archinect retorting: “No, Pokémon Go is not an urban fantasy for the new flaneur“. At Jacobin, Sam Kriss implores readers to “resist Pokémon Go“:
Walk around. Explore your neighborhood. Visit the park. Take in the sights. Have your fun. Pokémon Go is coercion, authority, a command issuing from out of a blank universe, which blasts through social and political cleavages to finally catch ‘em all. It must be resisted.
Some, like Jeff Sparrow at Overland, drew direct parallels to the Situationists:
On the one hand, that’s way cool – suddenly, the old pub near your house is inhabited by monsters.
On the other, there’s something faintly distasteful about the recuperation of specific real histories into a billion-dollar corporate mythology. Nearly 150 people lost their lives when the Triangle Shirtwaist Factory burned to the ground, entirely needless deaths caused by the atrocious working conditions of the garment trade. The tragedy became a rallying point for the trade union movement, the name of the factory, a shorthand reference to employers’ greed.
Now, though, it’s three free Pokeballs.
We might also say, then, that, even as the game leads players to embrace the derive, it also offers a remarkable demonstration of the phenomenon that Debord critiqued.
Writing for the Atlantic, Ian Bogost mediated on “the tragedy of Pokémon Go“:
We can have it both ways; we have to, even: Pokémon Go can be both a delightful new mechanism for urban and social discovery, and also a ghastly reminder that when it comes to culture, sequels rule. It’s easy to look at Pokémon Go and wonder if the game’s success might underwrite other, less trite or brazenly commercial examples of the genre. But that’s what the creators of pervasive games have been thinking for years, and still almost all of them are advertisements. Reality is and always has been augmented, it turns out. But not with video feeds of twenty-year old monsters in balls atop local landmarks. Rather, with swindlers shilling their wares to the everyfolk, whose ensuing dance of embrace and resistance is always as beautiful as it is ugly.
Pokémon Go’s popularity has led to many online comparisons to the Star Trek: TNG episode “The Game,” in which the crew of the Enterprise is overcome by a mind-controlling video game. The game in Star Trek is not strictly-speaking an augmented reality game, but does involve projecting images onto the player’s vision similar to an AR-overlay. Previous gaming and gadget fads have been compared to the TNG episode, notably Google Glass (for it’s similarity to the eye-beaming design used to interface with the game in Star Trek) and the pervasively popular Angry Birds game (as evident in this parody video). The comparison has regained cultural cachet because, unlike Angry Birds which can be played on the couch, Pokémon Go is played in motion. This, of course, has contributed to the perception of the game’s zombie-fying effects; we’ve grown accustomed to the fact that everyone’s eyes are glued to a smartphone screen in our public spaces, but now there are whole flocks of people milling around with their eyes on their devices.
My cynical side is inclined to agree with the critics who see Pokémon Go’s proliferation as proof positive of the passification and banalization of our society; the visions of Orwell, Bradbury, and Phil Dick all realized at once. But there’s something there that has me appreciative, even excited about this goofy game. As my wife and I wandered our neighborhood looking for pocket monsters, we noticed several other people walking around staring at their phones. This is not an uncommon sight, but it is re-contextualized in light of Pokémon Go’s popularity. “Look,” my wife would say, “I bet they’re playing, too.” After a while she had to know for sure, and started walking up to people and asking, “Are you playing Pokémon Go?” Every person she asked was indeed playing the game. Then we were walking along with these people we’ve just met, discussing play strategies, sharing Pokéstop locations, spreading word of upcoming lure parties.
One night around 10:30 last week we went into the Oakland neighborhood, home to both Pitt and Carnegie Mellon’s campuses and a hotbed of Pokémon Go activity. When we arrived, at least 20 people sat along the wall in front of the Soldiers & Sailors Memorial, smartphones in hands. We walked around the base of the Cathedral of Learning, where dozens of people in groups of two, three, or more were slowly pacing, stopping to capture a virtual creature. We crossed the street to Schenley plaza, where still dozens more people trekked through the grass, laughing and exclaiming and running up to their friends to share which Pokémon they had just got. Sure, most of these people were only talking to their own groups of friends, if they were talking at all, but it was still a cool experience. For me, the greatest thing was not which monsters I caught or XP my avatar earned; rather it was the energy, the unspoken but palpable buzz generated by all these people walking around in the dark of a warm summer night. Yes, I was giving attention to my smartphone screen, but what I remember most from that evening are the stars, and the fireflies, and the murmuring voices. Pokémon Go is promoting a sort of communal public activity, even if the sociality it produces is liminal at best. Yes, it is still shallow, still commercial, still programmed, but it’s something; there’s an energy there and a potential that is worth paying attention to.
Pokémon Go is not the be-all-end-all of augmented urban exploration, nor should be it considered the pinnacle of how mobile technology can enable new ways of interfacing with city space. But the game’s popularity, and my personal experience using it, has given me hope for the potential of AR apps to enrich our experience of urban spaces and engender new types of interactions in our shared environments.
- In Urban Media Ecology news, several recent studies reported correlations between characteristics of the built environment and human health. A study from the University of Kansas (in my birthplace of Lawrence) found that “neighborhoods that motivate walking can stave off cognitive decline in older adults“:
The researcher judged walkability using geographic information systems — essentially maps that measure and analyze spatial data.
“GIS data can tell us about roads, sidewalks, elevation, terrain, distances between locations and a variety of other pieces of information,” Watts said. “We then use a process called Space Syntax to measure these features, including the number of intersections, distances between places or connections between a person’s home and other possible destinations they might walk to. We’re also interested in how complicated a route is to get from one place to another. For example, is it a straight line from point A to point B, or does it require a lot of turns to get there?”
Watts said easy-to-walk communities resulted in better outcomes both for physical health—such as lower body mass and blood pressure—and cognition (such as better memory) in the 25 people with mild Alzheimer’s disease and 39 older adults without cognitive impairment she tracked. She believes that older adults, health care professionals, caregivers, architects and urban planners could benefit from the findings.
- Another research paper from researchers at multiple institutions “suggests that street design may have a larger impact on public health than previously thought“:
By studying 24 California cities with an array of street design characteristics and their associated health data, the authors find that living in cities with high intersection density—a measure of compactness—significantly reduces the risk of obesity, diabetes, high blood pressure, and heart disease. A full-grid street pattern also is a factor in lower risk of obesity, high blood pressure, and heart disease, as compared with full treelike patterns.
- At CityLab, Richard Florida posted a roundup of recent research studies on the health benefits afforded by walkable communities:
If walkability has long been an “ideal,” a recent slew of studies provide increasingly compelling evidence of the positive effects of walkable neighborhoods on everything from housing values to crime and health, to creativity and more democratic cities.
Walkability is no longer just an ideal. The evidence from a growing body of research shows that walkable neighborhoods not only raise housing prices but reduce crime, improve health, spur creativity, and encourage more civic engagement in our communities.
- Another CityLab article by Emily Von Hoffman looked at what architecture is doing to your brain:
I spoke with Dr. Julio Bermudez, the lead of a new study that uses fMRI to capture the effects of architecture on the brain. His team operates with the goal of using the scientific method to transform something opaque—the qualitative “phenomenologies of our built environment”—into neuroscientific observations that architects and city planners can deliberately design for. Bermudez and his team’s research question focuses on buildings and sites designed to elicit contemplation: They theorize that the presence of “contemplative architecture” in one’s environment may over time produce the same health benefits as traditional “internally induced” meditation, except with much less effort by the individual.
- As part of a directed study this semester, I’ve been studying the role of communication infrastructure in urban design, and particularly the parallel developments of mass media and the modern metropolis. Urban explorer Bradley Garrett recently wrote a piece for The Guardian about the massive infrastructure of underground London, including not just tube stations but communication infrastructure including Britain’s deepest telephone exchange:
The urban exploration crew I had worked with, the London Consolidation Crew or LCC, had long graduated from ruins and skyscrapers – it was the city in the city they were after, the secrets buried deep underground where the line between construction site and ruin is very thin indeed. The Kingsway Telephone Exchange was the crème de la crème, more coveted even than abandoned Tube stations or possibly even the forgotten Post Office railway we accessed in 2011.
Kingsway was originally built as a second world war air-raid shelter under Chancery Lane. These deep level shelters were, at one time, connected to the Tube and citizens would have undoubtedly taken refuge here during Luftwaffe bombing runs. In 1949 the tunnels were sold to the General Post Office where they became the termination for the first submarine transatlantic phone cable – the £120m TAT1 project. The system, meant to protect the vital connective tissue of the city in the event of terror-from-the-air (including nuclear attack), stretched for miles. It only had three surface entrances and contained a bar for workers on their off-hours, rumoured to be the deepest in the UK at 60m below the street. Although the government employed a host of people to maintain the tunnels, Kingsway was a spatial secret of state – part a trio of the most secure and sensitive telephone exchanges in Britain, along with the Anchor Exchange in Birmingham and the Guardian Exchange in Manchester.
- These photos of subterranean communication infrastructure contrast with images of above ground cables covering the urban landscape, as featured in this io9 post, “photos from the days when thousands of cables crowded the skies“:
Before most cables ran underground, all electrical, telephone and telegraph wires were suspended from high poles, creating strange and crowded streetscapes. Here are some typical views of late-19th century Boston, New York, Stockholm, and other wire-filled cities.
- Joseph Stromberg at Vox recently wrote about the “forgotten history of how automakers invented the crime of ‘jaywalking‘”:
Auto campaigners lobbied police to publicly shame transgressors by whistling or shouting at them — and even carrying women back to the sidewalk — instead of quietly reprimanding or fining them. They staged safety campaigns in which actors dressed in 19th century garb, or as clowns, were hired to cross the street illegally, signifying that the practice was outdated and foolish. In a 1924 New York safety campaign, a clown was marched in front of a slow-moving Model T and rammed repeatedly.
This strategy also explains the name that was given to crossing illegally on foot: jaywalking. During this era, the word “jay” meant something like “rube” or “hick” — a person from the sticks, who didn’t know how to behave in a city. So pro-auto groups promoted use of the word “jay walker” as someone who didn’t know how to walk in a city, threatening public safety.
- William Eric Rinehart at Sweet Talk asks: Did a change in rhetoric give rise to cities?
Between the mega-village and the cities that came later lies the formation of the state. Ultimately, this is the world of stratification buttressed through religion. With it came the creation of differing social groups and distinctions based upon rank or property. Yet, the acceptance of social specialization required a new view of the world, a new rhetoric in the McCloskeyian sense. And once that jump was made, benefits followed. Clustered people allowed for more trade and specialization of work, leading to more wealth, prestige and better equipped armies. While still a brutal world, cities had the potential for stability, but it came at the expense of radical equality.
- Matt Stroud at The Verge recently wrote about the dream and the myth of the paperless city in Chicago:
But you can’t just flip a switch to reverse paper systems in place for hundreds of years, can you? Adobe first released its Portable Document Format nearly 20 years ago, yet many private companies, nonprofit organizations, libraries, law firms, courts — and yes, major city governments such as Chicago’s — have yet to embrace a world reliant on PDFs and devoid of paper records. Mayor Emanuel has agreed to change that. Or at least to try. In 2011, he announced plans to spend $20 million on efficiency improvements including changes to make the city less reliant on paper.
Will Mayor Rahm Emanuel change the way governments deal with paper? Or is the road toward a “completely paperless” government a long way off?
- At URBACT, Stefan Höffken and Chris Haller consider urban planning and the multi-dimensional communication era:
“Because urban planning has always been based on the gathering and exchange of information and – as a democratic process – on communication between different stakeholders, a change in the method of communication has a significant impact on decision-making throughout the process”
The quote at the beginning of this post was taken out of a paper by Stefan Höffken and Chris Haller, who set out to research how new medias were used for urban planning matters. They are refering to geographer Manuel Castells’ and Clay Shirky‘s work to describe the change from uni-dimentional communication towards a many-to-many exchange sphere that, so Shirky, is on the verge of becoming ubiquitous. Höffken and Haller provide interesting insights in how different tools can serve certain goals and complement each-other by surveying urban projects and institutions or civil society mobilizations on urban matters as different as Tulsa municipality and the Mediaspree campaign in Berlin.
- Emily Badger at the Washington Post reports that Uber is offering cities an olive branch in the form of their customers’ trip data:
The company plans to partner first with Boston, sharing quarterly anonymized trip-level data with the city in a model that Uber says will become its national data-sharing policy. The data will include date, time, distance traveled and origin and destination locations for individual trips, identified only by zip code tabulation area to preserve privacy. Once held by cities, this information will be open to records requests, meaning that the public (and researchers) will have access to it, too.
Such data could help cities keep tabs on Uber and, for example, which neighborhoods the company is serving. Uber says, though, that it’s primarily offering the data so that cities can better understand themselves.
- Rosie Cima at priceonomics recently wrote about a designer’s war on misleading parking signs:
Sylianteng first tried to redesign parking signs when she was living in LA and applying to grad school, in a project she called “To Park or Not to Park.” She reduced the usual jumble of signs and regulations to a single, holistic panel, which looked a lot like a Google Calendar – it was a grid of days of the week, broken into hours. The blocks of time when a parking spot was valid she shaded green, the blocks of time it was invalid she shaded red. She also simplified the rules she illustrated, working off the principle that people would much rather adhere to an overly restrictive regulation than get a parking ticket.
Her prototypes provoked a lot of commentary, discussion, and praise. She used this feedback to improve her designs. She printed out new prototypes, and taped those up. The feedback validated some of her central assumptions, among them: (1) a lot of current parking signs were very confusing, and (2) people didn’t care why they could or couldn’t park somewhere, they just didn’t want to be ticketed.
- At Tropics of Meta, South El Monte Arts Posse posted about mapping community narratives in El Monte and South El Monte:
The writing of social history needs to keep in mind the motivations and individual agency of the people participating in events as they happen. In interview after interview, people were aware of the larger structural forces, and yet made choices and actions in contradiction to expectations. Again and again we spoke with people who beat the odds, who pushed back against racism, and took it upon themselves to change circumstances and in many cases succeeded.
- Arwa Mahdawi at The Guardian reports on neighborhood rebranding in London:
Similar semantic shifts are being attempted, with varying degrees of success, throughout the rest of London. Intrepid developers have discovered “Tyburnia”, an undervalued stretch of real estate between Paddington Station and Hyde Park. Meanwhile, the “Knowledge Quarter” is an attempt to rid King’s Cross of its association with prostitution by emphasising the new preponderance of cerebral institutions there. You could call it “brain-washing”. The Knowledge Quarter, incidentally, is one of 21 “Quarters” in London; there are also a dozen or so new “Villages”. Neighbourhood rebranding is often the linguistic leg of gentrification and, as such, follows a predictable pattern: “Villages” assert their legitimacy by emphasising community, while “Quarters” lend a gravitas to whatever noun they follow. Both have a cleansing effect on the associations that came before them.
- At NextCity, Nathan C. Martin considers why art, not Google, could revolutionize WiFi:
Remember a few years ago when television went digital and everyone had to get adapters or new TV sets? When that happened, what once were television channels became simply channels, a bulk of empty bandwidth that could host any variety of transmission. The Federal Communications Commission named it Super WiFi. The policies to regulate it are yet to be written, and a chorus is imploring the FCC to leave a large part of the spectrum open, or “unlicensed,” instead of auctioning it off. Those advocates tend to refer to the spectrum in spatial terms — a group of Stanford University economists likened the spectrum to a public park, a resource everyone should have access to. Mary Ellen Carroll speaks of it similarly. “It’s like public land,” she says. “It’s like Yosemite.”
- Finally, Emanuel Maiberg at Motherboard looks at homelessness in SimCity, and whether it is a bug or feature:
SimCity’s homeless people are represented as yellow, two-dimensional, ungendered figures with bags in tow. Their presence makes SimCity residents unhappy, and reduces land value. Like many other players, Bittanti discovered the online discussions when he was searching for a way to deal with them.
At first, players wondered if they were having so much problems with the homeless in their cities because of a bug in the game. Like many of 2014’s big-budget games that launched in broken or barely-functional states, SimCity originally would only work if players connected to EA’s servers, which repeatedly crashed under the load of players. It seemed possible that the homeless problem in SimCity was simply a mistake.
“Has anyone figured out a easy way to handle the homeless ruining those beautiful parks you spend so much money on?” asks one player on EA’s site. “Create jobs, either through zoning or upgrading road density near industry, that helped me a lot,” another player suggests.
- In an article for Haaretz reflecting on last week’s terror attacks in Paris, Michael Handelzalts invokes McLuhan’s infamous aphorism in relation to the emergence of print culture in the Islamic world:
So, in the Muslim world, books and literacy became generally accessible (instead of being accessible only to the educated male and the wealthy) about a quarter of a millennium later than in European-Western culture. I found this information, together with an assessment of the damage this 250-year lag caused to Muslim society and culture, in the works of Muslim scholars.
This lag could be made up in the blink of an eye as the cultural world moved from Johannes Gutenberg’s galaxy into the era when “The medium is the message,” and with the development of the virtual and digital world (at the expense of the printed one, of course).
- Today the Santa Barbara Independent published an article by Dean Stewart looking at McLuhan’s message 50 years after the publication of Understanding Media:
McLuhan had a lot of ideas and subsets of ideas. But he had one very big idea: that human civilization had passed through two stages of communication history, oral and print, and was embarking on another: electronic media. He believed the new media would change the way people relate to themselves and others and would change societies dramatically. Is the computer, then, the ubiquitous laptop and other devices, the McLuhan “audile-tactile” dream come true? There is no way to know. And it will take at least another 50 years to make a full evaluation of the work of Marshall McLuhan.
- At TechCrunch, Tadhg Kelly uses McLuhan’s famous formula to consider how mobile games are marketed:
Taking a leaf from McLuhan then, I submit that the message is the product. The tone, approach and strategy of how marketing is conducted shapes what kinds of product can be allowed by a product’s developer. What kind of ad you’ll run determines what kind of game you’ll believe can work, and therefore what kind of game you’ll fund and make.
The medium is the message and the message is the product, remember. In Marvel’s case the medium of cinema sends the message of the big experience, and the message disseminated through a high value trailer leads to the will to make a high value product: a big splashy movie. That’s how it earned the right to be thought of as premium. That’s how games do that too.
- At Newsday, Clarence Page uses McLuhan’s media theory to argue that the Internet can be used to undermine freedom:
When media guru Marshall McLuhan declared back in the 1960s that “Every innovation has within itself the seeds of its reversal,” I had no idea what he meant. But, like his other catchy quotables — “global village,” “cool media,” “the medium is the message” — it stayed with me. Now, in the Internet age, I am seeing proof of his prophecy every day.
For example, McLuhan predicted that a rapidly expanding automobile culture would lead to more traffic jams, air pollution and longing for space to take long walks or ride bicycles. I’m sure he’d give a knowing I-told-you-so nod to today’s battles between car people and bike people for asphalt space.
But more recently and less happily, I see far more sinister seeds of reversal in this era’s greatest innovation, the Internet. We greeted the Web as a liberator, but in today’s age of terrorism and post-Cold War autocrats it also poses a growing menace to the press freedoms it otherwise has invigorated.
- Lastly, Hervé St-Louis at ComicBookBin considers whether McLuhan is still relevant:
Two common critiques of McLuhan’s are his obliviousness to political economy and his technological determinism. McLuhan’s prognosis on media appears to celebrate a burgeoning world order and global capitalism. The way he foreshadows cognitive capitalism appears deterministic. Critics attack McLuhan for being silent on the transformation of global capitalism. This criticism focuses on what McLuhan did not write inUnderstanding Media as opposed to what he did. It is interesting to note that European scholars, even those who with political economic inclinations do not scorn McLuhan the way North Americans do. They do not blame him for being the messenger of a cognitive capitalist message.
McLuhan rightly described and to some extent predicted how messages need not be unidirectional. When he argued that technology is an extension of the senses, he did not argue that a select few had agency over the shaping of the message. He argued that any person had that potential. Specifically, he described how alternates modes of literacy allowed non literary people to participate in a global discourse. This is McLuhan’s legacy and part of why his work should be celebrated today.
- This EdTech article by D. Frank Smith showcases an app designed by a UW-Madison professor to teach campus sustainability through gamification:
One of the app’s games lets students practice waste sorting to reinforce good habits in distinguishing waste from recyclable materials. Different objects scroll down the side of the screen on a conveyer belt, and students click and drag to sort them into the proper bins.
Work on the app was funded in part by a Sustainability Innovation in Research and Education (SIRE) grant from the UW Office of Sustainability. It is being designed by the Mobile Learning Incubator (MLI) team at the Madison campus. Sustainable U is the team’s follow-up to Waste Eliminators, an interactive, environmental tour of campus which was also created with the help of Middlecamp and two graduate students in 2013, according to a university news release.
- Michael Logarta at GMA News reports on the expansion of homework-helping social platform Brainly to the Phillipines. The article touches on the value of ‘play’ for students, the challenges of gamifying education, and why playing is the future of learning:
“When lessons are delivered in a game-like manner – that is, interactive, challenging, with instantaneous feedback – students easily remember them,” said Christine Rom, Game Developers’ Association of the Philippines (GDAP) board member and CEO of educational games developer PODD. “If we harness the elements that make these games interesting and integrate them into education, we get the opportunity to enhance the learning process.”Gamification improves motivation. “By giving students a sense of achievement (through exciting rewards and recognition), students will be encouraged to learn and progress more,” explained Rom.
- Amanda Schneider at HuffPo considers gamification and the shift from “productivity” to “engagement”:
In a recent Mashable article Gabe Zichermann, the author of Game-Based Marketing and the CEO of Gamification.co states that one of the reasons for the increase in gamification is that long-standing marketing techniques are now failing. “They’re failing because people today are seeking more reward and more engagement from experiences than ever before.” Younger generations may be even more in tuned with gaming because they have likely never known a world without computers. Specifically, “Today’s youth mandates a more engaging experience. Gamification is required to bring those things into balance, and to make things engaging enough so people will pay attention to them and stay focused on them for a longer period of time.”
- Tim Biggs wrote an article at IGN considering video games in relation to the human search for authenticity:
The postmodern condition presents a constant struggle and conflict between our own desires and a world that seems fully available to experience but devoid of concrete or objective meaning. Video games, by virtue of their most basic structure, allow easy access to the feeling that your chosen actions and goals are both informed and legitimised by the overarching rules surrounding them. This is the very definition of authenticity.
None of this is to say that games are better or preferable experience to real life or other media, but it is to suggest they’re uniquely placed at this point in time to provide satisfying experiences. Indeed, matching the appeal of video games with the search for authenticity goes a ways to explaining the particular trajectory of gaming’s prevalence, from largely rejected as a toy in the production-focused late eighties and early nineties, to an explosion of mainstream acceptance as the global media and advertising machine makes up more and more of our everyday lives.
- Forbes contributor Michael Thomsen reports on an independent video game that touched on symbolic representation, labor and productivity, and even namedrops Franco Berardi’s “semiocapitalism”:
In most videogames, the semiotic meaning of the system is accepted by players before they begin playing—they don’t know what tactics they’ll use to win, nor whether they’ll play long enough to do so, but they know that winning or completion is the organizing metaphor. Players aren’t often encouraged to question the values of competitive systems, but only asked to internalize the responsibility of making them work as efficiently as possible, postponing the anxious reality of failure for a few magical moments that we’ve agreed to describe as fun.
In contrast, Rehearsals and Returns overflows with signifiers placed in a system that remains indifferent to their interpretative meaning, and which consciously obscures the player’s desire to interpret them in terms of winning or losing. The system acknowledges player choices—whether you chose to tell Hilary Clinton something hateful or nice—but the game doesn’t interpret the player’s choice, nor does it tie the economy of collectible conversation pieces to any allegorical meaning. It uses the game as a sort of digital confessional chamber, in which familiar units of social and political meaning are taken out of their historical narratives and given to the player in an incomplete space meant only for self-reflection.
- Ian Bogost recorded an interview for the Go For Rainbow podcast, discussing “gaming culture as it relates to geographical space, and when and when not to whip out the PhD cred“. Full audio is available here.
- David Harvey has a new book out, Seventeen Contradictions and the End of Capitalism. In this video Harvey speaks about the contradictions of capitalism (approx. 19 mins):
- I recently stumbled across this nearly hour-long documentary on Antonio Negri, titled The Revolt that Never Ends:
- Lastly, this video from Errant Signal considers Saints Row IV as celebration and example of kitsch (and makes me wish he’d dedicate a video to GTA V):
It’s been a long time since the last update (what happened to October?), so this post is extra long in an attempt to catch up.
- I haven’t seen the new Ender’s Game movie, but this review by abbeyotis at Cyborgology calls the film “a lean and contemporary plunge into questions of morality mediated by technology”:
In a world in which interplanetary conflicts play out on screens, the government needs commanders who will never shrug off their campaigns as merely “virtual.” These same commanders must feel the stakes of their simulated battles to be as high as actual warfare (because, of course, they are). Card’s book makes the nostalgic claim that children are useful because they are innocent. Hood’s movie leaves nostalgia by the roadside, making the more complex assertion that they are useful because of their unique socialization to be intimately involved with, rather than detached from, simulations.
- In the ongoing discourse about games criticism and its relation to film reviews, Bob Chipman’s latest Big Picture post uses his own review of the Ender’s Game film as an entry point for a breathless treatise on criticism. The video presents a concise and nuanced overview of arts criticism, from the classical era through film reviews as consumer reports up to the very much in-flux conceptions of games criticism. Personally I find this video sub-genre (where spoken content is crammed into a Tommy gun barrage of word bullets so that the narrator can convey a lot of information in a short running time) irritating and mostly worthless, since the verbal information is being presented faster than the listener can really process it. It reminds me of Film Crit Hulk, someone who writes excellent essays with obvious insight into filmmaking, but whose aesthetic choice (or “gimmick”) to write in all caps is often a distraction from the content and a deterrent to readers. Film Crit Hulk has of course addressed this issue and explained the rationale for this choice, but considering that his more recent articles have dropped the third-person “Hulk speak” writing style the all caps seems to be played out. Nevertheless, I’m sharing the video because Mr. Chipman makes a lot of interesting points, particularly regarding the cultural contexts for the various forms of criticism. Just remember to breathe deeply and monitor your heart rate while watching.
- In this video of a presentation titled Game design: the medium is the message, Jonathan Blow discusses how commercial constraints dictate the form of products from TV shows to video games.
- This somewhat related video from mynextappliance contextualizes Valve’s Steam machine place in gaming history.
- This video from Satchbag’s Goods is ostensibly a review of Hotline Miami, but develops into a discussion of art movements and Kanye West:
- This short interview with Slavoj Žižek in New York magazine continues a trend I’ve noticed since Pervert’s Guide to Ideology has been releasing, wherein writers interviewing Žižek feel compelled to include themselves and their reactions to/interactions with Žižek into their article. Something about a Žižek encounter brings out the gonzo in journalists. The NY mag piece is also notable for this succinct positioning of Žižek’s contribution to critical theory:
Žižek, after all, the Yugoslav-born, Ljubljana-based academic and Hegelian; mascot of the Occupy movement, critic of the Occupy movement; and former Slovenian presidential candidate, whose most infamous contribution to intellectual history remains his redefinition of ideology from a Marxist false consciousness to a Freudian-Lacanian projection of the unconscious. Translation: To Žižek, all politics—from communist to social-democratic—are formed not by deliberate principles of freedom, or equality, but by expressions of repressed desires—shame, guilt, sexual insecurity. We’re convinced we’re drawing conclusions from an interpretable world when we’re actually just suffering involuntary psychic fantasies.
- Wired UK reported on university students who turned maps of seventeenth century London into a detailed 3D world:
Following the development of the environment on the team’s blog you can see some of the gaps between what data was deemed noteworthy or worth recording in the seventeenth century and the level of detail we now expect in maps and other infographics. For example, the team struggled to pinpoint the exact location on Pudding Lane of the bakery where the Great Fire of London is thought to have originated and so just ended up placing it halfway along.
- Stephen Totilo reviewed the new pirate-themed Assassin’s Creed game for the New York Times. I haven’t played the game, but I love that the sections of the game set in the present day have shifted from the standard global conspiracy tropes seen in the earlier installments to postmodern self-referential and meta-fictional framing:
Curiously, a new character is emerging in the series: Ubisoft itself, presented mostly in the form of self-parody in the guise of a fictional video game company, Abstergo Entertainment. We can play small sections as a developer in Abstergo’s Montreal headquarters. Our job is to help turn Kenway’s life — mined through DNA-sniffing gadgetry — into a mass-market video game adventure. We can also read management’s emails. The team debates whether games of this type could sell well if they focused more on peaceful, uplifting moments of humanity. Conflict is needed, someone argues. Violence sells.
It turns out that Abstergo is also a front for the villainous Templars, who search for history’s secrets when not creating entertainment to numb the population. In these sections, Ubisoft almost too cheekily aligns itself with the bad guys and justifies its inevitable 2015 Assassin’s Creed, set during yet another violent moment in world history.
- Speaking of postmodern, self-referential, meta-fictional video games: The Stanley Parable was released late last month. There has already been a bevy of analysis written about the game, but I am waiting for the Mac release to play the game and doing my best to avoid spoilers in the meantime. Brenna Hillier’s post at VG24/7 is spoiler free (assuming you are at least familiar with the games premise, or its original incarnation as a Half Life mod), and calls The Stanley parable “a reaction against, commentary upon, critique and celebration of narrative-driven game design”:
The Stanley Parable wants you to think about it. The Stanley Parable, despite its very limited inputs (you can’t even jump, and very few objects are interactive) looks at those parts of first-person gaming that are least easy to design for – exploration and messing with the game’s engine – and foregrounds them. It takes the very limitations of traditional gaming narratives and uses them to ruthlessly expose their own flaws.
- An article at Techcrunch looks at how the Twitter-acquired Bluefin Labs “took the academic subject of semiotics and made it something “central” to the future of Twitter’s business“:
Roy’s research focus prior to founding Bluefin, and continued interest while running the company, has to do with how both artificial and human intelligences learn language. In studying this process, he determined that the most important factor in meaning making was the interaction between human beings: non one learns language in a vacuum, after all. That lesson helped inform his work at Twitter, which started with mapping the connection between social network activity and live broadcast television.
- Nathan at metopal posted their paper posing the question: What happens when we stop thinking about videogames as cinema and instead think of them through other media, like fashion, dance, or architecture?
Aspiring to cinematic qualities is not bad in an of itself, nor do I mean to shame fellow game writers, but developers and their attendant press tend to be myopic in their point of view, both figuratively and literally. If we continually view videogames through a monocular lens, we miss much of their potential. And moreover, we begin to use ‘cinematic’ reflexively without taking the time to explain what the hell that word means.
Metaphor is a powerful tool. Thinking videogames through other media can reframe our expectations of what games can do, challenge our design habits, and reconfigure our critical vocabularies. To crib a quote from Andy Warhol, we get ‘a new idea, a new look, a new sex, a new pair of underwear.’ And as I hinted before, it turns out that fashion and videogames have some uncanny similarities.
- John Powers at the Airship posted this great longform piece on the political economy of zombies:
Zombies started their life in the Hollywood of the 1930s and ‘40s as simplistic stand-ins for racist xenophobia. Post-millennial zombies have been hot-rodded by Danny Boyle and made into a subversive form of utopia. That grim utopianism was globalized by Max Brooks, and now Brad Pitt and his partners are working to transform it into a global franchise. But if zombies are to stay relevant, it will rely on the shambling monsters’ ability to stay subversive – and real subversive shocks and terror are not dystopian. They are utopian.
- This article at The Conversation addresses the “touchy subject” of Apple’s Touch ID:
Ironically, our bodies now must make physical contact with devices dictating access to the real; Apple’s Touch ID sensor can discern for the most part if we are actually alive. This way, we don’t end up trying to find our stolen fingers on the black market, or prevent others from 3D scanning them to gain access to our lives.
This is a monumental shift from when Apple released its first iPhone just six years ago. It’s a touchy subject: fingerprinting authentication means we confer our trust in an inanimate object to manage our animate selves – our biology is verified, digitised, encrypted, as they are handed over to our devices.
- In the wake of the Silk Road shut down last month, Chloe Albanesius at PC Mag asks: What was Silk Road and how did it work?
Can you really buy heroin on the Web as easily as you might purchase the latest best-seller from Amazon? Not exactly, but as the FBI explained in its complaint, it wasn’t exactly rocket science, thanks to Tor and some bitcoins. Here’s a rundown of how Silk Road worked before the feds swooped in.
- Henry Jenkins posted the transcript of an interview with Mark J.P. Wolf. The theme of the discussion is “imaginary worlds,” and they touch upon the narratology vs. ludology conflict in gaming:
The interactivity vs. storytelling debate is really a question of the author saying either “You choose” (interaction) or “I choose” (storytelling) regarding the events experienced; it can be all of one or all of the other, or some of each to varying degrees; and even when the author says “You choose”, you are still choosing from a set of options chosen by the author. So it’s not just a question of how many choices you make, but how many options there are per choice. Immersion, however, is a different issue, I think, which does not always rely on choice (such as immersive novels), unless you want to count “Continue reading” and “Stop reading” as two options you are constantly asked to choose between.
- Finally, GamesForChange has uploaded video of Ian Bogost’s keynote address from this year’s Games for Change Festival. Bogost extolls the virtues of “earnestness” over “seriousness” in game design:
- Kotaku recently posted a “manifesto” by game designer Eric Zimmerman declaring that the 21st century will be defined by games:
Systems, play, design: these are not just aspects of the Ludic Century, they are also elements of gaming literacy. Literacy is about creating and understanding meaning, which allows people to write (create) and read (understand).
New literacies, such as visual and technological literacy, have also been identified in recent decades. However, to be truly literate in the Ludic Century also requires gaming literacy. The rise of games in our culture is both cause and effect of gaming literacy in the Ludic Century.
- In a follow-up to Zimmerman’s treatise, Kotaku asked various game theorists, designers, and writers: will the 21st century be defined by games? Respondents include media theorist Ian Bogost:
So, perhaps there is one fundamental challenge for the Manifesto for a Ludic Century: would a truly ludic century be a century of manifestos? Of declaring simple principles rather than embracing systems? Or, is the Ludic Manifesto meant to be the last manifesto, the manifesto to end manifestos, replacing simple answers with the complexity of “information at play?”
- While on the topic of Bogost, I recently read his game studies paper from 2009 wherein he relates contemporary game studies to McLuhan’s discussion of games in 1964’s Understanding Media:
Might we conclude: videogames are the first creative medium to fully emerge after Marshall McLuhan. By the time they became popular, media ecology as a method was well-known. McLuhan was a popular icon. By the time the first generation of videogame players was becoming adults, McLuhan had become a trope. When the then-new publication Wired Magazine named him their “patron saint” in 1993, the editors didn’t even bother to explain what that meant. They didn’t need to.
By the time videogame studies became a going concern, McLuhan was gospel. So much so that we don’t even talk about him. To use McLuhan’s own language of the tetrad, game studies have enhanced or accelerated media ecology itself, to the point that the idea of studying the medium itself over its content has become a natural order.
- This piece by Bogost for The Atlantic looks at “flipped classrooms” and massive open online courses:
Generally speaking, educators have warmed to the idea of the flipped classroom far more than that of the MOOC. That move might be injudicious, as the two are intimately connected. It’s no accident that private, for-profit MOOC startups like Coursera have advocated for flipped classrooms, since those organizations have much to gain from their endorsement by universities. MOOCs rely on the short, video lecture as the backbone of a new educational beast, after all. Whether in the context of an all-online or a “hybrid” course, a flipped classroom takes the video lecture as a new standard for knowledge delivery and transfers that experience from the lecture hall to the laptop.
- Also, with increased awareness of Animal Crossing following from the latest game’s release for the Nintendo 3DS, Bogost recently posted an excerpt from his 2007 book Persuasive Games discussing consumption and naturalism in Animal Crossing:
Animal Crossing deploys a procedural rhetoric about the repetition of mundane work as a consequence of contemporary material property ideals. When my (then) five-year-old began playing the game seriously, he quickly recognized the dilemma he faced. On the one hand, he wanted to spend the money he had earned from collecting fruit and bugs on new furniture, carpets, and shirts. On the other hand, he wanted to pay off his house so he could get a bigger one like mine.
- In a post discussing the trio of protagonists of GTA V, in anticipation of the game’s impending release, Craig at unigamesity sees the use of multiple playable characters as addressing “ludonarrative dissonance“:
Ludonarrative dissonance is when the story the game is telling you and your gameplay experience somehow don’t match up. As an example, this was a particular issue in Rockstar’s most recent game, Max Payne 3. Max constantly makes remarks about how terrible he is at his job, even though he does more than is humanly possible to try to protect his employers – including making perfect one-handed head shots in mid-air while drunk and high on painkillers. The disparity and the dissonance between the narrative of the story and the gameplay leave things feeling off kilter and poorly inter-connnected. It doesn’t make sense or fit with your experience so it feels wrong and damages the cohesiveness of the game world and story. It’s like when you go on a old-lady only murdering spree as Niko, who is supposed to be a reluctant killer with a traumatic past, not a gerontophobic misogynist.
- Ludonarrative dissonance also cropped up in this post by Alex Pieschel on medium discussing irony and praise in videogames (found via Critical Distance):
What I find strange, in light of our supposed anti-irony cultural moment, is a kind of old-fashioned ironic conceit behind a number of recent critical darlings in the commercial videogame space. 2007’s Bioshock and this year’s Bioshock: Infinite are both about the irony of expecting ‘meaningful choice’ to live in an artificial dome of technological and commercial constraints. Last year’s Spec Ops: The Line offers a grim alchemy of self-deprecation and preemptive disdain for its audience. The Grand Theft Auto series has always maintained a cool, dismissive cynicism beneath its gleefully absurd mayhem. These games frame choice as illusory and experience as artificial. They are expensive, explosive parodies of free will.
- Which led me to this 2007 post from click nothing on ludonarrative dissonance in BioShock:
To cut straight to the heart of it, Bioshock seems to suffer from a powerful dissonance between what it is about as a game, and what it is about as a story. By throwing the narrative and ludic elements of the work into opposition, the game seems to openly mock the player for having believed in the fiction of the game at all. The leveraging of the game’s narrative structure against its ludic structure all but destroys the player’s ability to feel connected to either, forcing the player to either abandon the game in protest (which I almost did) or simply accept that the game cannot be enjoyed as both a game and a story, and to then finish it for the mere sake of finishing it.
- Sarah Wanenchak at Cyborgology considers mindfucks in videogames (partially in response to this post from Problem Machine):
The post itself makes a very important point: games, for the most part, can’t pull the Mindfuck like movies can because of the nature of the kind of storytelling to which most games are confined, which is predicated on a particular kind of interaction. Watching a movie may not be an entirely passive experience, but it’s clearly more passive than a game. You may identify with the characters on the screen, but you’re not meant to implicitly think of yourself as them. You’re not engaging in the kind of subtle roleplaying that most (mainstream) games encourage. You are not adopting an avatar. In a game, you are your profile, you are the character you create, and you are also to a certain degree the character that the game sets in front of you. I may be watching everything Lara Croft does from behind her, but I also control her; to the extent that she has choices, I make them. I get her from point A to B, and if she fails it’s my fault. When I talk about something that happened in the game, I don’t say that Lara did it. I say that I did.
- These excerpted research results published on kmjn.org reports that videogame players are acceptive of at least some cases of incoherence in anachronic gameplay:
Anachrony is a common storytelling technique in which events are narrated out of chronological order. A familiar example is a flashback, where story time jumps to the past for a bit, before returning to the present. The term “nonlinear narrative” is also sometimes used for this kind of out-of-order storytelling (somewhat less precisely).
While it’s a common technique in literature and film, anachrony is widely seen as more problematic to use in games, perhaps even to the point of being unusable. If the player’s actions during a flashback scene imply a future that differs considerably from the one already presented in a present-day scene (say, the player kills someone who they had been talking to in a present-day scene, or commits suicide in a flashback), this produces an inconsistent narrative. The root of the problem is that players generally have degree of freedom of action, so flashbacks are less like the case in literature and film—where already decided events are simply narrated out of order—and more like time travel, where the player travels back in time and can mess up the timeline.
- Polygon reports on a venture by Press Select to publish longform games criticism:
The first of the books are set to be published in early 2014. Some of the writers that will be published by Press Select in its first round have written for publications like Edge magazine, Kotaku, Kill Screen and personal blogs, including writers like Chris Dahlen, Michael Abbott, Jenn Frank, Jason Killingsworth, Maddy Myers, Tim Rogers, Patricia Hernandez and Robert Yang.
- I’ve long been fascinated by the gaming culture in South Korea, and Tom Massey has written a great feature piece for Eurogamer titled Seoul Caliber: Inside Korea’s Gaming Culture. From this westerner’s perspective, having never visited Korea, the article reads almost more like cyberpunk fiction than games journalism:
Not quite as ubiquitous, but still extremely common, are PC Bangs: LAN gaming hangouts where 1000 Won nets you an hour of multiplayer catharsis. In Gangnam’s Maxzone, overhead fans rotate at Apocalypse Now speed, slicing cigarette smoke as it snakes through the blades. Korea’s own NCSoft, whose European base is but a stone’s throw from the Eurogamer offices, is currently going strong with its latest MMO, Blade & Soul.
“It’s relaxing,” says Min-Su, sipping a Milkis purchased from the wall-mounted vending machine. “And dangerous,” he adds. “It’s easy to lose track of time playing these games, especially when you have so much invested in them. I’m always thinking about achieving the next level or taking on a quick quest to try to obtain a weapon, and the next thing I know I’ve been here for half the day.”
- As a cyberpunk/hyperreality aside, the city of Hong Kong has put up a blue-sky backdrop for when the real sky is too smoggy for tourist photos.
- In yet another cyberpunk dystopian tangent, I recently came across Chris Rogers’ site Fragments of a Hologram Rose: Re-seeing Blade Runner, with an assortment of content and analysis relating to the film.
- And one final cyberpunk diversion: this video is the first part of a lecture by University of Michigan professor Eric Rabkin covering cyberpunk, postmodernism, and beyond:
- Writing for The New Economy, Aaran Franda examines how the virtual economies seen in games like EVE Online provide valuable perspectives on real world economic activity:
Creation and simulation in virtual worlds appear to offer the best domain to test the new ideas required to tackle the very real problems of depravation, inequality, unemployment, and poverty that exist in national economies. On that note the need to see our socioeconomic institutions for the games that they really are seems even more poignant.
In the words of Vili Lehdonvirta, a leading scholar in virtual goods and currencies, the suffering we see today is “not some consequence of natural or physical law” it instead “is a result of the way we play these games.”
- Jon Evans at Tech Crunch looks at jobs, robots, capitalism, inequality, and you:
The global economy seems to be bifurcating into a rich/tech track and a poor/non-tech track, not least because new technology will increasingly destroy/replace old non-tech jobs. (Yes, global. Foxconn is already replacing Chinese employees with one million robots.) So far so fairly non-controversial.
The big thorny question is this: is technology destroying jobs faster than it creates them?
We live in an era of rapid exponential growth in technological capabilities. (Which may finally be slowing down, true, but that’s an issue for decades hence.) If you’re talking about the economic effects of technology in the 1980s, much less the 1930s or the nineteenth century, as if it has any relevance whatsoever to today’s situation, then you do not understand exponential growth. The present changes so much faster that the past is no guide at all; the difference is qualitative, not just quantitative. It’s like comparing a leisurely walk to relativistic speeds.
- This recent episode of Radiolab focused on talking to machines:
We begin with a love story–from a man who unwittingly fell in love with a chatbot on an online dating site. Then, we encounter a robot therapist whose inventor became so unnerved by its success that he pulled the plug. And we talk to the man who coded Cleverbot, a software program that learns from every new line of conversation it receives…and that’s chatting with more than 3 million humans each month. Then, five intrepid kids help us test a hypothesis about a toy designed to push our buttons, and play on our human empathy. And we meet a robot built to be so sentient that its creators hope it will one day have a consciousness, and a life, all its own.
- This video shows a demo of using Google Glass for interactive augmented reality:
- A recent Guardian article by Juliette Garside warns that our digital infrastructure is exceeding the limits of human control:
“These outages are absolutely going to continue,” said Neil MacDonald, a fellow at technology research firm Gartner. “There has been an explosion in data across all types of enterprises. The complexity of the systems created to support big data is beyond the understanding of a single person and they also fail in ways that are beyond the comprehension of a single person.”
From high volume securities trading to the explosion in social media and the online consumption of entertainment, the amount of data being carried globally over the private networks, such as stock exchanges, and the public internet is placing unprecedented strain on websites and on the networks that connect them.
- In an “anti-videogame manifesto,” Keith Burgun argues for intrinsic rewards and against grinding in videogames:
What I want is systems that have intrinsic rewards; that are disciplines similar to drawing or playing a musical instrument. I want systems which are their own reward.
What videogames almost always give me instead are labor that I must perform for an extrinsic reward. I want to convince you that not only is this not what I want, this isn’t really what anyone wants.
- This video from PBS Digital Studios’ Off Book looks at the rise of competetive gaming & e-sports:
- Will Luton at GamesIndustry International writes about the celebrification of game developers:
This ‘celebrification’ is enlivening making games and giving players role models, drawing more people in to development, especially indie and auteured games. This shift is proving more prosperous than any Skillset-accredited course or government pot could ever hope for. We are making men sitting in pants at their laptops for 12 hours a day as glamorous as it could be.
Creating luminaries will lead to all the benefits that more people in games can bring: a bigger and brighter community, plus new and fresh talent making exciting games. However, celebritydom demands storms, turmoil and gossip.
- The ongoing survey of Hollywood’s Summer of Doom continues with Isaac Chotiner’s New Republic article, Hollywood is in trouble and we’re all going to pay:
Spielberg’s theory is essentially that a studio will eventually go under after it releases five or six bombs in a row. The reason: budgets have become so gigantic. And, indeed, this summer has been full of movies with giant budgets and modest grosses, all of which has elicited hand-wringing about financial losses, the lack of a quality product (another post-apocalyptic thriller? more superheroes?), and a possible connection between the two. There has been some hope that Hollywood’s troubles will lead to a rethinking of how movies get made, and which movies get greenlit by studio executives. But a close look at this summer’s grosses suggest a more worrisome possibility: that the studios will become more conservative and even less creative.
- Finally, video of Slavoj Žižek and Paul A. Taylor discussing the difficulty of conveying philosophical ideas in today’s media:
- Media theorist and ludologist Ian Bogost recently penned some thoughts on Facebook’s development platform (referred to as “Facebook’s bleak new feudalism” in the title of Kotaku’s repost of the original piece):
The short truth is this: Facebook doesn’t care if developers can use the platform easily or at all. In fact, it doesn’t seem to concern itself with any of the factors that might be at play in developers’ professional or personal circumstances. The Facebook Platform is a selfish, self-made altar to Facebook, at which developers are expected to kneel and cower, rather than a generous contribution to the success of developers that also happens to benefit Facebook by its aggregate effects.
- Jeffrey van der Goot at Been Playing argues for “more Kubrickian and Lynchian narratives in video games“:
A lot of reactions to the narrative of [Bioshock] Infinite that I encountered were that it “didn’t make sense,” and that it was “being weird for the sake of being weird.”
Those reminded me of criticisms leveled at two of my favorite filmmakers: David Lynch and Stanley Kubrick. I think these comments arise because Infinite doesn’t go all the way, it hesitates. It tries to stick to conventional logic. It strews about Voxaphones to explain its abstractions.
- Shujaat Syed at Player Effort writes about “making linear story telling interesting in video games by acknowledging the fourth wall”:
At their core, video games are authoritarian. They have rules that need to be followed, and you are restricted to the game play systems and a story the programmers and designers have created. However, compared to other forms of media, they offer a breadth of freedom that is unmatched. I will not be speaking about the freedom of exploration. What I will be talking about is the freedom of creating a different type of narrative that is only possible through video games by breaking the 4th wall between the game and the player. This is one of our mediums greatest advantage, however, very rarely, is this power explored. With video games, we can have truly powerful forms of narrative, but at most we get ideas that could theoretically work as movies. Open-world sandbox games can dodge this because the player is free to create their own narrative alongside the main plotline, and this is a concept that is entirely unique to video games. It’s the linear story-based games where the narrative is usually much harder to distinguish than what you would get from a book or movie.
- Miles Klee at The Daily Dot reports on an app “that gamifies your boring sex life”:
In addition to registering your decibel levels (I’m hoping mine will get a boost from the garbage truck always idling outside my window), Spreadsheets will also monitor your overall duration, frequency, and somehow, thrusts per minute. Apparently this does not require supplementary electrodes.
What’s more, you can unlock “badges” and the like. For example, to meet the “Hello Sunshine” achievement, worth 10 points, you must take on the ultimate challenge of our time: “perform morning sex.”