• Flat Theory

    Flat Theory

    By David M. Berry
    ~

    The world is flat.[1] 6Or perhaps better, the world is increasingly “layers.” Certainly the augmediated imaginaries of the major technology companies are now structured around a post-retina vision of mediation made possible and informed by the digital transformations ushered in by mobile technologies – whether smartphones, wearables, beacons or nearables – an internet of places and things. These imaginaries provide a sense of place, as well as sense for management, of the complex real-time streams of information and data broken into shards and fragments of narrative, visual culture, social media and messaging. Turned into software, they reorder and re-present information, decisions and judgment, amplifying the sense and senses of (neoliberal) individuality whilst reconfiguring what it means to be a node in the network of post digital capitalism.  These new imaginaries serve as abstractions of abstractions, ideologies of ideologies, a prosthesis to create a sense of coherence and intelligibility in highly particulate computational capitalism (Berry 2014). To explore the experimentation of the programming industries in relation to this it is useful to explore the design thinking and material abstractions that are becoming hegemonic at the level of the interface.

    Two new competing computational interface paradigms are now deployed in the latest version of Apple and Google’s operating systems, but more notably as regulatory structures to guide the design and strategy related to corporate policy. The first is “flat design” which has been introduced by Apple through iOS 8 and OS X Yosemite as a refresh of the aging operating systems’ human computer interface guidelines, essentially stripping the operating system of historical baggage related to techniques of design that disguised the limitations of a previous generation of technology, both in terms of screen but also processor capacity. It is important to note, however, that Apple avoids talking about “flat design” as its design methodology, preferring to talk through its platforms specificity, that is about iOS’ design or OS X’s design. The second is “material design” which was introduced by Google into its Android L, now Lollipop, operating system and which also sought to bring some sense of coherence to a multiplicity of Android devices, interfaces, OEMs and design strategies. More generally “flat design” is “the term given to the style of design in which elements lose any type of stylistic characters that make them appear as though they lift off the page” (Turner 2014). As Apple argues, one should “reconsider visual indicators of physicality and realism” and think of the user interface as “play[ing] a supporting role”, that is that techniques of mediation through the user interface should aim to provide a new kind of computational realism that presents “content” as ontologically prior to, or separate from its container in the interface (Apple 2014). This is in contrast to “rich design,” which has been described as “adding design ornaments such as bevels, reflections, drop shadows, and gradients” (Turner 2014).

    color_family_a_2xI want to explore these two main paradigms – and to a lesser extent the flat-design methodology represented in Windows 7 and the, since renamed, Metro interface – through a notion of a comprehensive attempt by both Apple and Google to produce a rich and diverse umwelt, or ecology, linked through what what Apple calls “aesthetic integrity” (Apple 2014). This is both a response to their growing landscape of devices, platforms, systems, apps and policies, but also to provide some sense of operational strategy in relation to computational imaginaries. Essentially, both approaches share an axiomatic approach to conceptualizing the building of a system of thought, in other words, a primitivist predisposition which draws from both a neo-Euclidian model of geons (for Apple), but also a notion of intrinsic value or neo-materialist formulations of essential characteristics (for Google). That is, they encapsulate a version of what I am calling here flat theory. Both of these companies are trying to deal with the problematic of multiplicities in computation, and the requirement that multiple data streams, notifications and practices have to be combined and managed within the limited geography of the screen. In other words, both approaches attempt to create what we might call aggregate interfaces by combining techniques of layout, montage and collage onto computational surfaces (Berry 2014: 70).

    The “flat turn” has not happened in a vacuum, however, and is the result of a new generation of computational hardware, smart silicon design and retina screen technologies. This was driven in large part by the mobile device revolution which has not only transformed the taken-for-granted assumptions of historical computer interface design paradigms (e.g. WIMP) but also the subject position of the user, particularly structured through the Xerox/Apple notion of single-click functional design of the interface. Indeed, one of the striking features of the new paradigm of flat design, is that it is a design philosophy about multiplicity and multi-event. The flat turn is therefore about modulation, not about enclosure, as such, indeed it is a truly processual form that constantly shifts and changes, and in many ways acts as a signpost for the future interfaces of real-time algorithmic and adaptive surfaces and experiences. The structure of control for the flat design interfaces is following that of the control society, is “short-term and [with] rapid rates of turnover, but also continuous and without limit” (Deleuze 1992). To paraphrase Deleuze: Humans are no longer in enclosures, certainly, but everywhere humans are in layers.

    manipulation_2x

    Apple uses a series of concepts to link its notion of flat design which include, aesthetic integrity, consistency, direct manipulation, feedback, metaphors, and user control (Apple 2014). Reinforcing the haptic experience of this new flat user interface has been described as building on the experience of “touching glass” to develop the “first post-Retina (Display) UI (user interface)” (Cava 2013). This is the notion of layered transparency, or better, layers of glass upon which the interface elements are painted through a logical internal structure of Z-axis layers. This laminate structure enables meaning to be conveyed through the organization of the Z-axis, both in terms of content, but also to place it within a process or the user interface system itself.

    Google, similarly, has reorganized it computational imaginary around a flattened layered paradigm of representation through the notion of material design. Matias Duarte, Google’s Vice President of Design and a Chilean computer interface designer, declared that this approach uses the notion that it “is a sufficiently advanced form of paper as to be indistinguishable from magic” (Bohn 2014). But magic which has constraints and affordances built into it, “if there were no constraints, it’s not design — it’s art” Google claims (see Interactive Material Design) (Bohn 2014). Indeed, Google argues that the “material metaphor is the unifying theory of a rationalized space and a system of motion”, further arguing:

    The fundamentals of light, surface, and movement are key to conveying how objects move, interact, and exist in space and in relation to each other. Realistic lighting shows seams, divides space, and indicates moving parts… Motion respects and reinforces the user as the prime mover… [and together] They create hierarchy, meaning, and focus (Google 2014).

    This notion of materiality is a weird materiality in as much as Google “steadfastly refuse to name the new fictional material, a decision that simultaneously gives them more flexibility and adds a level of metaphysical mysticism to the substance. That’s also important because while this material follows some physical rules, it doesn’t create the “trap” of skeuomorphism. The material isn’t a one-to-one imitation of physical paper, but instead it’s ‘magical’” (Bohn 2014). Google emphasises this connection, arguing that “in material design, every pixel drawn by an application resides on a sheet of paper. Paper has a flat background color and can be sized to serve a variety of purposes. A typical layout is composed of multiple sheets of paper” (Google Layout, 2014). The stress on material affordances, paper for Google and glass for Apple are crucial to understanding their respective stances in relation to flat design philosophy.[2]

    • Glass (Apple): Translucency, transparency, opaqueness, limpidity and pellucidity.
    • Paper (Google): Opaque, cards, slides, surfaces, tangibility, texture, lighted, casting shadows.
    Paradigmatic Substances for Materiality

    In contrast to the layers of glass paper-notes-templatethat inform the logics of transparency, opaqueness and translucency of Apple’s flat design, Google uses the notion of remediated “paper” as a digital material, that is this “material environment is a 3D space, which means all objects have x, y, and z dimensions. The z-axis is perpendicularly aligned to the plane of the display, with the positive z-axis extending towards the viewer.  Every sheet of material occupies a single position along the z-axis and has a standard 1dp thickness” (Google 2014). One might think then of Apple as painting on layers of glass, and Google as thin paper objects (material) placed upon background paper. However a key difference lies in the use of light and shadow in Google’s notion which enables the light source, located in a similar position to the user of the interface, to cast shadows of the material objects onto the objects and sheets of paper that lie beneath them (see Jitkoff 2014). Nonetheless, a laminate structure is key to the representational grammar that constitutes both of these platforms.

    armin_hofmann_2
    Armin Hofmann, head of the graphic design department at the Schule für Gestaltung Basel (Basel School of Design) and was instrumental in developing the graphic design style known as the Swiss Style. Designs from 1958 and 1959.

    Interestingly, both design strategies emerge from an engagement with and reconfiguration of the principles of design that draw from the Swiss style (sometimes called the International Typographic Style) in design (Ashghar 2014, Turner 2014).[3] This approach emerged in the 1940s, and

    mainly focused on the use of grids, sans-serif typography, and clean hierarchy of content and layout. During the 40’s and 50’s, Swiss design often included a combination of a very large photograph with simple and minimal typography (Turner 2014).

    The design grammar of the Swiss style has been combined with minimalism and the principle of “responsive design”, that is that the materiality and specificity of the device should be responsive to the interface and context being displayed. Minimalism is a “term used in the 20th century, in particular from the 1960s, to describe a style characterized by an impersonal austerity, plain geometric configurations and industrially processed materials” (MoMA 2014).

    img-robert-morris-1_125225955286
    Robert Morris: Untitled (Scatter Piece), 1968-69, felt, steel, lead, zinc, copper, aluminum, brass, dimensions variable; at Leo Castelli Gallery, New York. Photo Genevieve Hanson. All works © 2010 Robert Morris/Artists Rights Society (ARS), New York.

    Robert Morris, one of the principle artists of Minimalism, and author of the influential Notes on Sculpture used “simple, regular and irregular polyhedrons. Influenced by theories in psychology and phenomenology” which he argued “established in the mind of the beholder ‘strong gestalt sensation’, whereby form and shape could be grasped intuitively” (MoMA 2014).[4]

    The implications of these two competing world-views are far-reaching in that much of the worlds initial contact, or touch points, for data services, real-time streams and computational power is increasingly through the platforms controlled by these two companies. However, they are also deeply influential across the programming industries, and we see alternatives and multiple reconfigurations in relation to the challenge raised by the “flattened” design paradigms. That is, they both represent, if only in potentia, a situation of a power relation and through this an ideological veneer on computation more generally. Further, with the proliferation of computational devices – and the screenic imaginary associated with them in the contemporary computational condition – there appears a new logic which lies behind, justifies and legitimates these design methodologies.

    It seems to me that these new flat design philosophies, in the broad sense, produce an order in precepts and concepts in order to give meaning and purpose not only in the interactions with computational platforms, but also more widely in terms of everyday life. Flat design and material design are competing philosophies that offer alternative patterns of both creation and interpretation, which are meant to have not only interface design implications, but more broadly in the ordering of concepts and ideas, the practices and the experience of computational technologies broadly conceived. Another way to put this could be to think about these moves as being a computational founding, the generation of, or argument for, an axial framework for building, reconfiguration and preservation.

    Indeed, flat design provides and more importantly serves, as a translational or metaphorical heuristic for both re-presenting the computational, but also teaches consumers and users how to use and manipulate new complex computational systems and stacks. In other words, in a striking visual technique flat design communicates the vertical structure of the computational stack, on which the Stack corporations are themselves constituted. But also begins to move beyond the specificity of the device as privileged site of a computational interface interaction from beginning to end. For example, interface techniques are abstracted away from the specificity of the device, for example through Apple’s “handoff” continuity framework which also potentially changes reading and writing practices in interesting ways and new use-cases for wearables and nearables.

    These new interface paradigms, introduced by the flat turn, have very interesting possibilities for the application of interface criticism, through unpacking and exploring the major trends and practices of the Stacks, that is, the major technology companies. I think that further than this, the notion of layers are instrumental in mediating the experience of an increasingly algorithmic society (e.g. think dashboards, personal information systems, quantified self, etc.), and as such provide an interpretative frame for a world of computational patterns but also a constituting grammar for building these systems in the first place. There is an element in which the notion of the postdigital may also be a useful way into thinking about the question of the link between art, computation and design given here (see Berry and Dieter, forthcoming) but also the importance of notions of materiality for the conceptualization deployed by designers working within both the flat design and material design paradigms – whether of paper, glass, or some other “material” substance.[5]
    _____

    David M. Berry is Reader in the School of Media, Film and Music at the University of Sussex. He writes widely on computation and the digital and blogs at Stunlaw. He is the author of Critical Theory and the Digital, The Philosophy of Software: Code and Mediation in the Digital Age , Copy, Rip, Burn: The Politics of Copyleft and Open Source, editor of Understanding Digital Humanities and co-editor of Postdigital Aesthetics: Art, Computation And Design. He is also a Director of the Sussex Humanities Lab.

    Back to the essay
    _____

    Notes

    [1] Many thanks to Michael Dieter and Søren Pold for the discussion which inspired this post.

    [2] The choice of paper and glass as the founding metaphors for the flat design philosophies of Google and Apple raise interesting questions for the way in which these companies articulate the remediation of other media forms, such as books, magazines, newspapers, music, television and film, etc. Indeed, the very idea of “publication” and the material carrier for the notion of publication is informed by the materiality, even if only a notional affordance given by this conceptualization. It would be interesting to see how the book is remediated through each of the design philosophies that inform both companies, for example.

    [3] One is struck by the posters produced in the Swiss style which date to the 1950s and 60s but which today remind one of the mobile device screens of the 21st Century.

    [4] There is also some interesting links to be explored between the Superflat style and postmodern art movement, founded by the artist Takashi Murakami, which is influenced by manga and anime, both in terms of the aesthetic but also in relation to the cultural moment in which “flatness” is linked to “shallow emptiness.”

    [5] There is some interesting work to be done in thinking about the non-visual aspects of flat theory, such as the increasing use of APIs, such as the RESTful api, but also sound interfaces that use “flat” sound to indicate spatiality in terms of interface or interaction design. There are also interesting implications for the design thinking implicit in the Apple Watch, and the Virtual Reality and Augmented Reality platforms of Oculus Rift, Microsoft HoloLens, Meta and Magic Leap.

    Bibliography
  • What Melts in the "Melting Pot" of Hong Kong?

    Umbella Man

    an essay by
    Hong Kong Baptist University professor and boundary 2 advisory editor
    Ruth Y.Y. Hung
    ~

    “This paper analyses the vicissitudes of Hong Kong people‘s waves of hostility toward visitors from mainland China, treating the hostility as exemplary of a more general but extensive problem of racialism in China.”

    Click here to read the entire essay published by Asiatic.

  • What Drives Automation?

    What Drives Automation?

    glass-cagea review of Nicholas Carr, The Glass Cage: Automation and Us (W.W. Norton, 2014)
    by Mike Bulajewski
    ~

    Debates about digital technology are often presented in terms of stark polar opposites: on one side, cyber-utopians who champion the new and the cutting edge, and on the other, cyber-skeptics who hold on to obsolete technology. The framing is one-dimensional in the general sense that it is superficial, but also in a more precise and mathematical sense that it implicitly treats the development of technology as linear. Relative to the present, there are only two possible positions and two possible directions to move; one can be either for or against, ahead or behind.[1]

    Although often invoked as a prelude to transcending the division and offering a balanced assessment, in describing the dispute in these pro or con terms one has already betrayed one’s orientation, tilting the field against the critical voice by assigning it an untenable position. Criticizing a new technology is misconstrued as a simple defense of the old technology or of no technology, which turns legitimate criticism into mere conservative fustiness, a refusal to adapt and a failure to accept change.

    Few critics of technology match these descriptions, and those who do, like the anarcho-primitivists who claim to be horrified by contemporary technology, nonetheless accede to the basic framework set by technological apologists. The two sides disagree only on the preferred direction of travel, making this brand of criticism more pro-technology than it first appears. One should not forget that the high-tech futurism of Silicon Valley is supplemented by the balancing counterweight of countercultural primitivism, with Burning Man expeditions, technology-free Waldorf schools for children of tech workers, spouses who embrace premodern New Age beliefs, romantic agrarianism, and restorative digital detox retreats featuring yoga and meditation. The diametric opposition between pro- and anti-technology is internal to the technology industry, perhaps a symptom of the repression of genuine debate about the merits of its products.

    ***

    Nicholas Carr’s most recent book, The Glass Cage: Automation and Us, is a critique of the use of automation and a warning of its human consequences, but to conclude, as some reviewers have, that he is against automation or against technology as such is to fall prey to this one-dimensional fallacy.[2]

    The book considers the use of automation in areas like medicine, architecture, finance, manufacturing and law, but it begins with an example that’s closer to home for most of us: driving a car. Transportation and wayfinding are minor themes throughout the book, and with Google and large automobile manufacturers promising to put self-driving cars on the street within a decade, the impact of automation in this area may soon be felt in our daily lives like never before. Early in the book, we are introduced to problems that human factors engineers working with airline autopilot systems have discovered and may be forewarnings of a future of the unchecked automating of transportation.

    Carr discusses automation bias—the tendency for operators to assume the system is correct and external signals that contradict it are wrong—and the closely related problem of automation complacency, which occurs when operators assume the system is infallible and so abandon their supervisory role. These problems have been linked to major air disasters and are behind less-catastrophic events like oblivious drivers blindly following their navigation systems into nearby lakes or down flights of stairs.

    The chapter dedicated to deskilling is certain to raise the ire of skeptical readers because it begins with an account of the negative impact of GPS technology on Inuit hunters who live in the remote northern reaches of Canada. As GPS devices proliferated, hunters lost what a tribal elder describes as “the wisdom and knowledge of the Inuit”: premodern wayfinding methods that rely on natural phenomena like wind, stars, tides and snowdrifts to navigate. Inuit wayfinding skills are truly impressive. The anthropologist Claudio Aporta reports traveling with a hunter across twenty square kilometers of flat featureless land as he located seven fox traps that he had never seen before, set by his uncle twenty five years prior. These talents have been eroded as Inuit hunters have adopted GPS devices that seem to do the job equally well, but have the unexpected side effect of increasing injuries and deaths as hunters succumb to equipment malfunctions and the twin perils of automation complacency and bias.

    Laboring under the misconceptions of the one-dimensional fallacy, it would be natural to take this as a smoking gun of Carr’s alleged anti-technology perspective and privileging of the premodern, but the closing sentences of the chapter point us away from that conclusion:

    We ignore the ways that software programs and automated systems might be reconfigured so as not to weaken our grasp on the world but to strengthen it. For, as human factors researchers and other experts on automation have found, there are ways to break the glass cage without losing the many benefits computers grant us. (151)

    These words segue into the following chapter, where Carr identifies the dominant philosophy that designs automation technologies to inadvertently produce problems that he identified earlier: technology-centered automation. This approach to design is distrustful of humans, perhaps even misanthropic. It views us as weak, inefficient, unreliable and error-prone, and seeks to minimize our involvement in the work to be done. It institutes a division of labor between human and machine that gives the bulk of the work over to the machine, only seeking human input in anomalous situations. This philosophy is behind modern autopilot systems that hand off control to human pilots for only a few minutes in a flight.

    The fundamental argument of the book is that this design philosophy can lead to undesirable consequences. Carr seeks an alternative he calls human-centered automation, an approach that ensures the human operator remains engaged and alert. Autopilot systems designed with this philosophy might return manual control to the pilots at irregular intervals to ensure they remain vigilant and practice their flying skills. It could provide tactile feedback of its operations so that pilots are involved in a visceral way rather than passively monitoring screens. Decision support systems like those used in healthcare could take a secondary role of reviewing and critiquing a decision made by a doctor made rather than the other way around.

    The Glass Cage calls for a fundamental shift in how we understand error. Under the current regime, an error is an inefficiency or an inconvenience, to be avoided at all costs. As defined by Carr, a human-centered approach to design treats error differently, viewing it as an opportunity for learning. He illustrates this with a personal experience of repeatedly failing a difficult mission in the video game Red Dead Redemption, and points to the satisfaction of finally winning a difficult game as an example of what is lost when technology is designed to be too easy. He offers video games as a model for the kinds of technologies he would like to see: tools that engage us in difficult challenges, that encourage us to develop expertise and to experience flow states.

    But Carr has an idiosyncratic definition of human-centered design which becomes apparent when he counterposes his position against the prominent design consultant Peter Merholz. Echoing premises almost universally adopted by designers, Merholz calls for simple, frictionless interfaces and devices that don’t require a great deal of skill, memorization or effort to operate. Carr objects that that eliminates learning, skill building and mental engagement—perhaps a valid criticism, but it’s strange to suggest that this reflects a misanthropic technology-centered approach.

    A frequently invoked maxim of human-centered design is that technology should adapt to people, rather than people adapting to technology. In practice, the primary consideration is helping the user achieve his or her goal as efficiently and effectively as possible, removing unnecessary obstacles and delays that stand in the way. Carr argues for the value of challenges, difficulties and demands placed on users to learn and hone skills, all of which fall under the prohibited category of people adapting to technology.

    In his example of playing Red Dead Redemption, Carr prizes the repeated failure and frustration before finally succeeding at the game. From the lens of human-centered design, that kind of experience is seen as a very serious problem that should be eliminated quickly, which is probably why this kind of design is rarely employed at game studios. In fact, it doesn’t really make sense to think of a game player as having a goal, at least not from the traditional human-centered standpoint. The driver of a car has a goal: to get from point A to point B; a Facebook user wants to share pictures with friends; the user of a word processor wants to write a document; and so on. As designers, we want to make these tasks easy, efficient and frictionless. The most obvious way of framing game play is to say that the player’s goal is to complete the game. We would then proceed to remove all obstacles, frustrations, challenges and opportunities for error that stand in the way so that they may accomplish this goal more efficiently, and then there would be nothing left for them to do. We would have ruined the game.

    This is not necessarily the result of a misanthropic preference for technology over humanity, though it may be. It is also the likely outcome of a perfectly sincere and humanistic belief that we shouldn’t inconvenience the user with difficulties that stand in the way of their goal. As human factors researcher David Woods puts it, “The road to technology-centered systems is paved with human-centered intentions,”[3] a phrasing which suggests that these two philosophies aren’t quite as distinct as Carr would have us believe.

    Carr’s vision of human-centered design differs markedly from contemporary design practice, which stresses convenience, simplicity, efficiency for the user and ease of use. In calling for less simplicity and convenience, he is in effect critical of really existing human-centeredness, and that troubles any reading of The Glass Cage that views it a book about restoring our humanity in a world driven mad by machines.

    It might be better described as a book about restoring one conception of humanity in a world driven mad by another. It is possible to argue that the difference between the two appears in psychoanalytic theory as the difference between drive and desire. The user engages with a technology in order to achieve a goal because they perceive themselves as lacking something. Through the use of this tool, they believe they can regain it and fill in this lack. It follows that designers ought to help the user achieve their goal—to reach their desire—as quickly and efficiently as possible because this will satisfy them and make them happy.

    But the insight of psychoanalysis is that lack is ontological and irreducible, it cannot be filled in any permanent way because any concrete lack we experience is in fact metonymic for a constitutive lack of being. As a result, as desiring subjects we are caught in an endless loop of seeking out that object of desire, feeling disappointed when we find it because it didn’t fulfill our fantasies and then finding a new object to chase. The alternative is to shift from desire to drive, turning this failure into a triumph. Slavoj Žižek describes drive as follows: “the very failure to reach its goal, the repetition of this failure, the endless circulation around the object, generates a satisfaction of its own.”[4]

    This satisfaction is perhaps what Carr aims at when he celebrates the frustrations and challenges of video games and of work in general. That video games can’t be made more efficient without ruining them indicates that what players really want is for their goal to be thwarted, evoking the psychoanalytic maxim that summarizes the difference between desire and drive: from the missing/lost object, to loss itself as an object. This point is by no means tangential. Early on, Carr introduces the concept of miswanting, defined as the tendency to desire what we don’t really want and won’t make us happy—in this case, leisure and ease over work and challenge. Psychoanalysts holds that all human wanting (within the register of desire) is miswanting. Through fantasy, we imagine an illusory fullness or completeness of which actual experience always falls short.[5]

    Carr’s revaluation of challenge, effort and, ultimately, dissatisfaction cannot represent a correction of the error of miswanting­–of rediscovering the true source of pleasure and happiness in work. Instead, it radicalizes the error: we should learn to derive a kind of satisfaction from our failure to enjoy. Or, in the final chapter, as Carr says of the farmer in Robert Frost’s poem Mowing, who is hard at work and yet far from the demands of productivity: “He’s not seeking some greater truth beyond the work. The work is the truth.”

    ***

    Nicholas Carr has a track record of provoking designers to rethink their assumptions. With The Shallows, along with other authors making related arguments, he influenced software developers to create a new class of tools that cut off the internet, eliminate notifications or block social media web sites to help us concentrate. Starting with OS X Lion in 2011, Apple began offering a full screen mode that hides distracting interface elements and background windows from inactive applications.

    What transformative effects could The Glass Cage have on the way software is designed? The book certainly offers compelling reasons to question whether ease of use should always be paramount. Advocates for simplicity are rarely challenged, but they may now find themselves facing unexpected objections. Software could become more challenging and difficult to use—not in the sense of a recalcitrant WiFi router that emits incomprehensible error codes, but more like a game. Designers might draw inspiration from video games, perhaps looking to classics like the first level of Super Mario Brothers, a masterpiece of level design that teaches the fundamental rules of the game without ever requiring the player to read the manual or step through a tutorial.

    Everywhere that automation now reigns, new possibilities announce themselves. A spell checker might stop to teach spelling rules, or make a game out of letting the user take a shot at correcting mistakes it has detected. What if there was a GPS navigation device that enhanced our sense of spatial awareness rather than eroding it, that engaged our attention on to the road rather than let us tune out. Could we build an app that helps drivers maintain good their skills by challenging them to adopt safer and more fuel-efficient driving techniques?

    Carr points out that the preference for easy-to-use technologies that reduce users’ engagement is partly a consequence of economic interests and cost reduction policies that profit from the deskilling and reduction of the workforce, and these aren’t dislodged simply by pressing for new design philosophies. But to his credit, Carr has written two best-selling books aimed at the general interest reader on the fairly obscure topic of human-computer interaction. User experience designers working in the technology industry often face an uphill battle in trying to build human-centered products (however that is defined). When these matters attract public attention and debate, it makes their job a little easier.

    _____

    Mike Bulajewski (@mrteacup) is a user experience designer with a Master’s degree from University of Washington’s Human Centered Design and Engineering program. He writes about technology, psychoanalysis, philosophy, design, ideology & Slavoj Žižek at MrTeacup.org. He has previously written about the Spike Jonze film Her for The b2 Review Digital Studies section.

    Back to the essay

    _____

    [1] Differences between individual technologies are ignored and replaced by the monolithic master category of Technology. Jonah Lehrer’s review of Nicholas Carr’s 2010 book The Shallows in the New York Times exemplifies such thinking. Lehrer finds contradictory evidence against Carr’s argument that the internet is weakening our mental faculties in scientific studies that attribute cognitive improvements to playing video games, a non sequitur which gains meaning only by subsuming these two very different technologies under a single general category of Technology. Evgeny Morozov is one of the sharpest critics of this tendency. Here one is reminded of his retort in his article “Ghosts in the Machine” (2013): “That dentistry has been beneficial to civilization tells us very little about the wonders of data-mining.”

    [2] There are a range of possible causes for this constrictive linear geometry: a tendency to see a progressive narrative of history; a consumerist notion of agency which only allows shoppers to either upgrade or stick with what they have; or the oft-cited binary logic of digital technology. One may speculate about the influence of the popular technology marketing book by Geoffrey A. Moore, Crossing the Chasm (2014) whose titular chasm is the gap between the elite group of innovators and early adopters—the avant-garde—and the recalcitrant masses bringing up the rear who must be persuaded to sign on to their vision.

    [3] David D. Woods and David Tinapple (1999). “W3: Watching Human Factors Watch People at Work.” Proceedings of the 43rd Annual Meeting of the Human Factors and Ergonomics Society (1999).

    [4] Slavoj Žižek, The Parallax View (2006), 63.

    [5] The cultural and political implications of this shift are explored at length in Todd McGowan’s two books The End of Dissatisfaction: Jacques Lacan and the Emerging Society of Enjoyment (2003) and Enjoying What We Don’t Have: The Political Project of Psychoanalysis (2013).

  • Thinking the Permissible, or Speaking in Common

    Thinking the Permissible, or Speaking in Common

    _____

    In this paper written for the MLA roundtable “Toward an Academic Commons: Academic Freedom and Sites of Contested Speech,” Colin Dayan reflects on what happens when academic freedom is cornered and subsumed by the demands of “civility” and “rationality.”

    _____

    And these are knaves that brawl for better laws
    And cant of tyranny in stronger powers
    Who glut their vile unsatiated maws
    And freedoms birthright from the weak devour

    –John Clare, “To a Fallen Elm”

    Who gets to speak? “To speak,” Fanon wrote in Black Skin, White Masks is to exist for the other. “To speak means to be in a position to use a certain syntax, to grasp the morphology of this or that language, but it means above all to assume a culture, to support the weight of a civilization.” To acquire a “civilizing language,” then, is to be an exile in your own land, however we define the terrain of personal significance and conviction.

    The “academic commons,” is something of an oxymoron, a paradox as heady as St. Paul’s encounter with the vexation of bodily resurrection. Must not a new definition of the body be coined? The crux of the matter lies in the spirit that is also the flesh, in something quite intangible, even insensate, but persistent and always quite thing-like.

    The call to the commons, to a “common ground,” or, as some would endorse “the common interests,” is to revel in what we as teachers and students share “in common.” But who is to say what the ground of the common, the commonplace or commonality is? As Morrison wrote in Beloved, “definitions are always in the hands of the definers.” To hold ideas in common is much like the call for consensus: whoever calls for this fellowship has a pretty good idea of what is or is not to be accepted or permissible as shared values. Or to put it more bluntly, these values are akin to holding certain “truths to be self-evident.” As Justice Taney made explicit in his opinion in Dred Scott: “we the people” was never meant to include that class of persons who carried the “indelible mark” of “degradation”:

    The men who framed this declaration were great men — high in literary acquirements, high in their sense of honor, and incapable of asserting principles inconsistent with those on which they were acting. They perfectly understood the meaning of the language they used, and how it would be understood by others, and they knew that it would not in any part of the civilized world be supposed to embrace the negro race, which, by common consent, had been excluded from civilized Governments and the family of nations, and doomed to slavery.

    Which people, or persons, or members of the global Academy are excluded from the commons? Zones of exclusion mark off indelibly the lives that are cordoned off, trapped, demolished, bulldozed, expropriated, uprooted, by-passed.

    But for some of our colleagues, those of us who would bring these lives to light are guilty of “partisan politics,” “purges,” and “censorship,” even though there is absolutely no evidence of such practice or intent. Politics is never separate from scholarship. Thought, imagination, and learning is always enhanced by collision and conflict. In “On Lyric Poetry and Society,” Theodore Adorno wrote, and I paraphrase, any poet who claims not to be political is in fact proclaiming a politics—and I would add—silencing others in a far more devious because subtle way.

    * * * * * * * * *

    Why should I speak? I’ve always been haunted by Gramsci’s warning that we are all “experts in legitimation” and by the savvy insight of Bourdieu and Passeron in Reproduction in Education, Society, and Culture published in 1977, that we – no matter what we intend—can never shed that ermine robe of authority. An authority honed to say what needs to be taught and thought in the right, the tasteful, the distinctive and recognizable—the civil—way.

    But there is always a politics to the under-read and the critical background against which such silences are made. William Carlos Williams recognized the silencing generated by experts, the careening authority of the estimable critics. Before the formidable and raging Book 3 of Paterson, against the background of a library burning and books in flames, he scorned the “knowledgeable idiots, the university,/…../The outward masks of the special interests/that perpetuate the stasis and make it profitable.”
    He recognized how inextricable is the language, the craft, the tools of scholarship from the casual disregard, the everyday violence in the “corrupt cities.” So it was not artlessness that he wanted but something else, something he called “the foulness”—the kind of writing that threatened the moneyed and the privileged, the extensions of state-sanctioned racism and its sanguine exclusivity.

    The affront to the reasonable, the necessary, and the secure matters now more than ever. Reasonable consensus and civility: these words engage me, unsettled as I am by the prospect of divisions (not only of subject but also genre) that allow the continued dispossession of those creatures – human and non-human alike–who remain outside the circle of grace, delivered to subjection without recourse. How can we shed the mantle of civility, consensus, and rationality just long enough to question the claims of decency?

    How do we speak? When it comes to matters of civility and befouling, we need to question our role as teachers and scholars, as well the threats to our profession by a silencing that is racially driven though culturally masked. The remedy of culture always excuses or conceals the experience of racism. In this time of terror, in our complicity with “kill lists,” drone strikes, global dispossession, mass incarceration, and murders of black citizens on our streets, what are the choices offered us as academics?

    A cure for all kinds of threats, reasonableness has long been a way to extend persecution, civil death, and torture. But this rationality, like the theory that accompanies it, is tied to figurative power, and its metaphors can at any time become more insistent. What the anthropologist and historian, the late Michel-Rolph Trouillot, called the “explanatory power of culture” allowed the contexts for inequality and racism to continue. Indeed, they became harsher because hidden by the call for a common ground.

    Our embrace of the “commons” or “academic commons,” though well-intentioned, risks enabling the false ideal of “consensus” that always rears its ugly, if “reasonable” head just when social and racial stratification is at its worst. Against the fashionable cartography of the commons, or a search for common ground, stands the language of stigma, incarceration, control—and downright extermination. Should we not reclaim the singularity of lives that do not span borders, or more precisely, who never —at least not in this country—gained the right to have rights in common?

    Matters of terminology delimit privilege, just as they silence the disenfranchised, the invisible ones, who are always quite visible though objects of serene disregard. Our thought should be supple and ever watchful for the terms that perpetuate the very contexts of inequality—and specifically, racism. I’m grappling here with the push-and-pull of the call for the commons, always moderated and even reproduced by the humanitarian concern that is analogous to it, and hence always already closed to criticism. I learned from Trouillot—may he rest in peace—how the most benign of academic trends carried with it a strategy of extermination that always targets people of color, the poor or the powerless. He knew its eugenic violence even as he confronted its universalizing promise, what he called “totalitarian humanism” and the bugbears of academia: civility and compromise.

    * * * * * * * * *

    Retaliation hides behind a logic that is at once exclusionary and authoritarian. This policing of thought indulges in all manner of intolerant behavior. In the context of the de-hiring of a scholar like Salaita, we have sought in our conscience how to respond to coercion and threat found in the least likely of places—the university’s so-called “ivory tower.”

    Sickened by the civility-mongers’ pieties and the costs to scholars who dare to think hard and feel passionately, I recall Poe’s disdain for the oracles of “higher morality.” These “thinkers-that-they-think” used “civility”–and all such words for self-righteous sentiment–to destroy literary careers and exclude anyone who threatened the status quo. Such entrenched proclivities for what Poe condemned as “doggerel aesthetics” masked the ugly money-fundament to such cunning ideality. What has happened to Steven Salaita matters to us all, no matter our views or our assumed status in the groves of academe. The resilience of genuine academic freedom is that it ensures that these abuses, and the watchdog groups and alumni that abet them, are no longer hidden from the eyes of the world, immune to the prescriptive force of morality, beyond the judgment of society, masked by the appeal to “civility.”

  • Frank Pasquale — To Replace or Respect: Futurology as if People Mattered

    Frank Pasquale — To Replace or Respect: Futurology as if People Mattered

    a review of Erik Brynjolfsson and Andrew McAfee, The Second Machine Age: Work, Progress, and Prosperity in a Time of Brilliant Technologies (W.W. Norton, 2014)

    by Frank Pasquale

    ~

    Business futurism is a grim discipline. Workers must either adapt to the new economic realities, or be replaced by software. There is a “race between education and technology,” as two of Harvard’s most liberal economists insist. Managers should replace labor with machines that require neither breaks nor sick leave. Superstar talents can win outsize rewards in the new digital economy, as they now enjoy global reach, but they will replace thousands or millions of also-rans. Whatever can be automated, will be, as competitive pressures make fairly paid labor a luxury.

    Thankfully, Erik Brynjolfsson and Andrew McAfee’s The Second Machine Age (2MA)  downplays these zero-sum tropes. Brynjolffson & McAfee (B&M) argue that the question of distribution of the gains from automation is just as important as the competitions for dominance it accelerates. 2MA invites readers to consider how societies will decide what type of bounty from automation they want, and what is wanted first.  The standard, supposedly neutral economic response (“whatever the people demand, via consumer sovereignty”) is unconvincing. As inequality accelerates, the top 5% (of income earners) do 35% of the consumption. The top 1% is responsible for an even more disproportionate share of investment. Its richest members can just as easily decide to accelerate the automation of the wealth defense industry as they can allocate money to robotic construction, transportation, or mining.

    A humane agenda for automation would prioritize innovations that complement (jobs that ought to be) fulfilling vocations, and substitute machines for dangerous or degrading work. Robotic meat-cutters make sense; robot day care is something to be far more cautious about. Most importantly, retarding automation that controls, stigmatizes, and cheats innocent people, or sets up arms races with zero productive gains, should be a much bigger part of public discussions of the role of machines and software in ordering human affairs.

    2MA may set the stage for such a human-centered automation agenda. Its diagnosis of the problem of rapid automation (described in Part I below) is compelling. Its normative principles (II) are eclectic and often humane. But its policy vision (III) is not up to the challenge of channeling and sequencing automation. This review offers an alternative, while acknowledging the prescience and insight of B&M’s work.

    I. Automation’s Discontents

    For B&M, the acceleration of automation ranks with the development of agriculture, or the industrial revolution, as one of the “big stories” of human history (10-12). They offer an account of the “bounty and spread” to come from automation. “Bounty” refers to the increasing “volume, variety, and velocity” of any imaginable service or good, thanks to its digital reproduction or simulation (via, say, 3-D printing or robots). “Spread” is “ever-bigger differences among people in economic success” that they believe to be just as much an “economic consequence” of automation as bounty.[1]

    2MA briskly describes various human workers recently replaced by computers.  The poor souls who once penned corporate earnings reports for newspapers? Some are now replaced by Narrative Science, which seamlessly integrates new data into ready-made templates (35). Concierges should watch out for Siri (65). Forecasters of all kinds (weather, home sales, stock prices) are being shoved aside by the verdicts of “big data” (68). “Quirky,” a startup, raised $90 million by splitting the work of making products between a “crowd” that “votes on submissions, conducts research, suggest improvements, names and brands products, and drives sales” (87), and Quirky itself, which “handles engineering, manufacturing, and distribution.” 3D printing might even disintermediate firms like Quirky (36).

    In short, 2MA presents a kaleidoscope of automation realities and opportunities. B&M skillfully describe the many ways automation both increases the “size of the pie,” economically, and concentrates the resulting bounty among the talented, the lucky, and the ruthless. B&M emphasize that automation is creeping up the value chain, potentially substituting machines for workers paid better than the average.

    What’s missing from the book are the new wave of conflicts that would arise if those at very top of the value chain (or, less charitably, the rent and tribute chain) were to be replaced by robots and algorithms. When BART workers went on strike, Silicon Valley worthies threatened to replace them with robots. But one could just as easily call for the venture capitalists to be replaced with algorithms. Indeed, one venture capital firm added an algorithm to its board in 2013.  Travis Kalanick, the CEO of Uber, responded to a question on driver wage demands by bringing up the prospect of robotic drivers. But given Uber’s multiple legal and PR fails in 2014, a robot would probably would have done a better job running the company than Kalanick.

    That’s not “crazy talk” of communistic visions along the lines of Marx’s “expropriate the expropriators,” or Chile’s failed Cybersyn.[2]  Thiel Fellow and computer programming prodigy Vitaly Bukherin has stated that automation of the top management functions at firms like Uber and AirBnB would be “trivially easy.”[3] Automating the automators may sound like a fantasy, but it is a natural outgrowth of mantras (e.g., “maximize shareholder value”) that are commonplaces among the corporate elite. To attract and retain the support of investors, a firm must obtain certain results, and the short-run paths to attaining them (such as cutting wages, or financial engineering) are increasingly narrow.  And in today’s investment environment of rampant short-termism, the short is often the only term there is.

    In the long run, a secure firm can tolerate experiments. Little wonder, then, that the largest firm at the cutting edge of automation—Google—has a secure near-monopoly in search advertising in numerous markets. As Peter Thiel points out in his recent From Zero to One, today’s capitalism rewards the best monopolist, not the best competitor. Indeed, even the Department of Justice’s Antitrust Division appeared to agree with Thiel in its 1995 guidelines on antitrust enforcement in innovation markets. It viewed intellectual property as a good monopoly, the rightful reward to innovators for developing a uniquely effective process or product. And its partner in federal antitrust enforcement, the Federal Trade Commission, has been remarkably quiescent in response to emerging data monopolies.

    II. Propertizing Data

    For B&M, intellectual property—or, at least, the returns accruing to intellectual insight or labor—plays a critical role in legitimating inequalities arising out of advanced technologies.  They argue that “in the future, ideas will be the real scarce inputs in the world—scarcer than both labor and capital—and the few who provide good ideas will reap huge rewards.”[4] But many of the leading examples of profitable automation are not “ideas” per se, or even particularly ingenious algorithms. They are brute force feats of pattern recognition: for example, Google’s studying past patterns of clicks to see what search results, and what ads, are personalized to delight and persuade each of its hundreds of millions of users. The critical advantage there is the data, not the skill in working with it.[5] Google will demur, but if they were really confident, they’d license the data to other firms, confident that others couldn’t best their algorithmic prowess.  They don’t, because the data is their critical, self-reinforcing advantage. It is a commonplace in big data literatures to say that the more data one has, the more valuable any piece of it becomes—something Googlers would agree with, as long as antitrust authorities aren’t within earshot.

    As sensors become more powerful and ubiquitous, feats of automated service provision and manufacture become more easily imaginable.  The Baxter robot, for example, merely needs to have a trainer show it how to move in order to ape the trainer’s own job. (One is reminded of the stories of US workers flying to India to train their replacements how to do their job, back in the day when outsourcing was the threat du jour to U.S. living standards.)

    how to train a robot
    How to train a Baxter robot. Image source: Inc. 

    From direct physical interaction with a robot, it is a short step to, say, programmed holographic or data-driven programming.  For example, a surveillance camera on a worker could, after a period of days, months, or years, potentially record every movement or statement of the worker, and replicate it, in response to whatever stimuli led to the prior movements or statements of the worker.

    B&M appear to assume that such data will be owned by the corporations that monitor their own workers.  For example, McDonalds could train a camera on every cook and cashier, then download the contents into robotic replicas. But it’s just as easy to imagine a legal regime where, say, workers’ rights to the data describing their movements would be their property, and firms would need to negotiate to purchase the rights to it.  If dance movements can be copyrighted, so too can the sweeps and wipes of a janitor. Consider, too, that the extraordinary advances in translation accomplished by programs like Google Translate are in part based on translations by humans of United Nations’ documents released into the public domain.[6] Had the translators’ work not been covered by “work-made-for-hire” or similar doctrines, they might well have kept their copyrights, and shared in the bounty now enjoyed by Google.[7]

    Of course, the creativity of translation may be greater than that displayed by a janitor or cashier. Copyright purists might thus reason that the merger doctrine denies copyrightability to the one best way (or small suite of ways) of doing something, since the idea of the movement and its expression cannot be separated. Grant that, and one could still imagine privacy laws giving workers the right to negotiate over how, and how pervasively, they are watched. There are myriad legal regimes governing, in minute detail, how information flows and who has control over it.

    I do not mean to appropriate here Jaron Lanier’s ideas about micropayments, promising as they may be in areas like music or journalism. A CEO could find some critical mass of stockers or cooks or cashiers to mimic even if those at 99% of stores demanded royalties for the work (of) being watched. But the flexibility of legal regimes of credit, control, and compensation is under-recognized. Living in a world where employers can simply record everything their employees do, or Google can simply copy every website that fails to adopt “robots.txt” protection, is not inevitable. Indeed, according to renowned intellectual property scholar Oren Bracha, Google had to “stand copyright on its head” to win that default.[8]

    Thus B&M are wise to acknowledge the contestability of value in the contemporary economy.  For example, they build on the work of MIT economists Daron Acemoglu and David Autor to demonstrate that “skill biased technical change” is a misleading moniker for trends in wage levels.  The “tasks that machines can do better than humans” are not always “low-skill” ones (139). There is a fair amount of play in the joints in the sequencing of automation: sometimes highly skilled workers get replaced before those with a less complex and difficult-to-learn repertoire of abilities.  B&M also show that the bounty predictably achieved via automation could compensate the “losers” (of jobs or other functions in society) in the transition to a more fully computerized society. By seriously considering the possibility of a basic income (232), they evince a moral sensibility light years ahead of the “devil-take-the-hindmost” school of cyberlibertarianism.

    III. Proposals for Reform

    Unfortunately, some of B&M’s other ideas for addressing the possibility of mass unemployment in the wake of automation are less than convincing.  They praise platforms like Lyft for providing new opportunities for work (244), perhaps forgetting that, earlier in the book, they described the imminent arrival of the self-driving car (14-15). Of course, one can imagine decades of tiered driving, where the wealthy get self-driving cars first, and car-less masses turn to the scrambling drivers of Uber and Lyft to catch rides. But such a future seems more likely to end in a deflationary spiral than  sustainable growth and equitable distribution of purchasing power. Like the generation traumatized by the Great Depression, millions subjected to reverse auctions for their labor power, forced to price themselves ever lower to beat back the bids of the technologically unemployed, are not going to be in a mood to spend. Learned helplessness, retrenchment, and miserliness are just as likely a consequence as buoyant “re-skilling” and self-reinvention.

    Thus B&M’s optimism about what they call the “peer economy” of platform-arranged production is unconvincing.  A premier platform of digital labor matching—Amazon’s Mechanical Turk—has occasionally driven down the wage for “human intelligence tasks” to a penny each. Scholars like Trebor Scholz and Miriam Cherry have discussed the sociological and legal implications of platforms that try to disclaim all responsibility for labor law or other regulations. Lilly Irani’s important review of 2MA shows just how corrosive platform capitalism has become. “With workers hidden in the technology, programmers can treat [them] like bits of code and continue to think of themselves as builders, not managers,” she observes in a cutting aside on the self-image of many “maker” enthusiasts.

    The “sharing economy” is a glidepath to precarity, accelerating the same fate for labor in general as “music sharing services” sealed for most musicians. The lived experience of many “TaskRabbits,” which B&M boast about using to make charts for their book, cautions against reliance on disintermediation as a key to opportunity in the new digital economy. Sarah Kessler describes making $1.94 an hour labeling images for a researcher who put the task for bid on Mturk.  The median active TaskRabbit in her neighborhood made $120 a week; Kessler cleared $11 an hour on her best day.

    Resistance is building, and may create fairer terms online.  For example, Irani has helped develop a “Turkopticon” to help Turkers rate and rank employers on the site. Both Scholz and Mike Konczal have proposed worker cooperatives as feasible alternatives to Uber, offering drivers both a fairer share of revenues, and more say in their conditions of work. But for now, the peer economy, as organized by Silicon Valley and start-ups, is not an encouraging alternative to traditional employment. It may, in fact, be worse.

    Therefore, I hope B&M are serious when they say “Wild Ideas [are] Welcomed” (245), and mention the following:

    • Provide vouchers for basic necessities. . . .
    • Create a national mutual fund distributing the ownership of capital widely and perhaps inalienably, providing a dividend stream to all citizens and assuring the capital returns do not become too highly concentrated.
    • Depression-era Civilian Conservation Corps to clean up the environment, build infrastructure.

    Speaking of the non-automatable, we could add the Works Progress Administration (WPA) to the CCC suggestion above.  Revalue the arts properly, and the transition may even add to GDP.

    Soyer, Artists on the WPA
    Moses Soyer, “Artists on WPA” (1935). Image source: Smithsonian American Art Museum

    Unfortunately, B&M distance themselves from the ideas, saying, “we include them not necessarily to endorse them, but instead to spur further thinking about what kinds of interventions will be necessary as machines continue to race ahead” (246).  That is problematic, on at least two levels.

    First, a sophisticated discussion of capital should be at the core of an account of automation,  not its periphery. The authors are right to call for greater investment in education, infrastructure, and basic services, but they need a more sophisticated account of how that is to be arranged in an era when capital is extraordinarily concentrated, its owners have power over the political process, and most show little to no interest in long-term investment in the skills and abilities of the 99%. Even the purchasing power of the vast majority of consumers is of little import to those who can live off lightly taxed capital gains.

    Second, assuming that “machines continue to race ahead” is a dodge, a refusal to name the responsible parties running the machines.  Someone is designing and purchasing algorithms and robots. Illah Reza Nourbaksh’s Robot Futures suggests another metaphor:

    Today most nonspecialists have little say in charting the role that robots will play in our lives.  We are simply watching a new version of Star Wars scripted by research and business interests in real time, except that this script will become our actual world. . . . Familiar devices will become more aware, more interactive and more proactive; and entirely new robot creatures will share our spaces, public and private, physical and digital. . . .Eventually, we will need to read what they write, we will have to interact with them to conduct our business transactions, and we will often mediate our friendships through them.  We will even compete with them in sports, at jobs, and in business. [9]

    Nourbaksh nudges us closer to the truth, focusing on the competitive angle. But the “we” he describes is also inaccurate. There is a group that will never have to “compete” with robots at jobs or in business—rentiers. Too many of them are narrowly focused on how quickly they can replace needy workers with undemanding machines.

    For the rest of us, another question concerning automation is more appropriate: how much can we be stuck with? A black-card-toting bigshot will get the white glove treatment from AmEx; the rest are shunted into automated phone trees. An algorithm determines the shifts of retail and restaurant workers, oblivious to their needs for rest, a living wage, or time with their families.  Automated security guards, police, and prison guards are on the horizon. And for many of the “expelled,” the homines sacres, automation is a matter of life and death: drone technology can keep small planes on their tracks for hours, days, months—as long as it takes to execute orders.

    B&M focus on “brilliant technologies,” rather than the brutal or bumbling instances of automation.  It is fun to imagine a souped-up Roomba making the drudgery of housecleaning a thing of the past.  But domestic robots have been around since 2000, and the median wage-earner in the U.S. does not appear to be on a fast track to a Jetsons-style life of ease.[10] They are just as likely to be targeted by the algorithms of the everyday, as they are to be helped by them. Mysterious scoring systems routinely stigmatize persons, without them even knowing. They reflect the dark side of automation—and we are in the dark about them, given the protections that trade secrecy law affords their developers.

    IV. Conclusion

    Debates about robots and the workers “struggling to keep up” with them are becoming stereotyped and stale. There is the standard economic narrative of “skill-biased technical change,” which acts more as a tautological, post hoc, retrodictive, just-so story than a coherent explanation of how wages are actually shifting. There is cyberlibertarian cornucopianism, as Google’s Ray Kurzweil and Eric Schmidt promise there is nothing to fear from an automated future. There is dystopianism, whether intended as a self-preventing prophecy, or entertainment. Each side tends to talk past the other, taking for granted assumptions and values that its putative interlocutors reject out of hand.

    Set amidst this grim field, 2MA is a clear advance. B&M are attuned to possibilities for the near and far future, and write about each in accessible and insightful ways.  The authors of The Second Machine Age claim even more for it, billing it as a guide to epochal change in our economy. But it is better understood as the kind of “big idea” book that can name a social problem, underscore its magnitude, and still dodge the elaboration of solutions controversial enough to scare off celebrity blurbers.

    One of 2MA’s blurbers, Clayton Christensen, offers a backhanded compliment that exposes the core weakness of the book. “[L]earners and teachers alike are in a perpetual mode of catching up with what is possible. [The Second Machine Age] frames a future that is genuinely exciting!” gushes Christensen, eager to fold automation into his grand theory of disruption. Such a future may be exciting for someone like Christensen, a millionaire many times over who won’t lack for food, medical care, or housing if his forays fail. But most people do not want to be in “perpetually catching up” mode. They want secure and stable employment, a roof over their heads, decent health care and schooling, and some other accoutrements of middle class life. Meaning is found outside the economic sphere.

    Automation could help stabilize and cheapen the supply of necessities, giving more persons the time and space to enjoy pursuits of their own choosing. Or it could accelerate arms races of various kinds: for money, political power, armaments, spying, stock trading. As long as purchasing power alone—whether of persons or corporations—drives the scope and pace of automation, there is little hope that the “brilliant technologies” B&M describe will reliably lighten burdens that the average person experiences. They may just as easily entrench already great divides.

    All too often, the automation literature is focused on replacing humans, rather than respecting their hopes, duties, and aspirations. A central task of educators, managers, and business leaders should be finding ways to complement a workforce’s existing skills, rather than sweeping that workforce aside. That does not simply mean creating workers with skill sets that better “plug into” the needs of machines, but also, doing the opposite: creating machines that better enhance and respect the abilities and needs of workers.  That would be a “machine age” welcoming for all, rather than one calibrated to reflect and extend the power of machine owners.

    _____

    Frank Pasquale (@FrankPasquale) is a Professor of Law at the University of Maryland Carey School of Law. His recent book, The Black Box Society: The Secret Algorithms that Control Money and Information (Harvard University Press, 2015), develops a social theory of reputation, search, and finance.  He blogs regularly at Concurring Opinions. He has received a commission from Triple Canopy to write and present on the political economy of automation. He is a member of the Council for Big Data, Ethics, and Society, and an Affiliate Fellow of Yale Law School’s Information Society Project. He is a frequent contributor to The b2 Review Digital Studies section.

    Back to the essay
    _____

    [1] One can quibble with the idea of automation as necessarily entailing “bounty”—as Yves Smith has repeatedly demonstrated, computer systems can just as easily “crapify” a process once managed well by humans. Nor is “spread” a necessary consequence of automation; well-distributed tools could well counteract it. It is merely a predictable consequence, given current finance and business norms and laws.

    [2] For a definition of “crazy talk,” see Neil Postman, Stupid Talk, Crazy Talk: How We Defeat Ourselves by the Way We Talk and What to Do About It (Delacorte, 1976). For Postman, “stupid talk” can be corrected via facts, whereas “crazy talk” “establishes different purposes and functions than the ones we normally expect.” If we accept the premise of labor as a cost to be minimized, what better to cut than the compensation of the highest paid persons?

    [3] Conversation with Sam Frank at the Swiss Institute, Dec. 16, 2014, sponsored by Triple Canopy.

    [4] In Brynjolfsson, McAfee, and Michael Spence, “New World Order: Labor, Capital, and Ideas in the Power Law Economy,” an article promoting the book. Unfortunately, as with most statements in this vein, B&M&S give us little idea how to identify a “good idea” other than one that “reap[s] huge rewards”—a tautology all too common in economic and business writing.

    [5] Frank Pasquale, The Black Box Society (Harvard University Press, 2015).

    [6] Programs, both in the sense of particular software regimes, and the program of human and technical efforts to collect and analyze the translations that were the critical data enabling the writing of the software programs behind Google Translate.

    [9] Illah Reza Nourbaksh, Robot Futures (MIT Press, 2013), pp. xix-xx.

    [10] Erwin Prassler and Kazuhiro Kosuge, “Domestic Robotics,” in Bruno Siciliano and Oussama Khatib, eds., Springer Handbook of Robotics (Springer, 2008), p. 1258.

  • Letter from Warsaw

    Letter from Warsaw

    POLIN

    an essay by Charles Bernstein
    ~
    In November 2014, my wife, Susan Bee, and I visited Lodz and Warsaw, Poland, for the first time, along with our friend, the poet Tracie Morris. Susan’s mother was born in Lodz, moving to Berlin when she was around five. When I spoke at the American centers at the University of Lodz and Warsaw, I acknowledged standing on the ground of a great experiment from the decade before I was born, which aimed to expel difference in order to increase harmony through sameness, but the move toward homogenization versus miscegenation is powerful in the Americas as well, since the poetics of the Americas is the continuation of European poetics by other means.

    When I explore the poetics of representation of the Systematic Extermination Process, I said in Warsaw, I don’t do so as an American looking at Poland and Germany but as someone whose intellectual and cultural foundations are European. I feel it as much my story as anyone else now alive and that a part of European culture destroyed here lives on with me, in and as an American, and is expressed through a commitment to the syncretic and miscegenated poetics of Americas.

    The aversion of an originary or authentic or correct language is foundational for the poetics of Americas and makes a sharp contrast with those European (and American) nationalists that place a single language as fundamental to national or literary identity or who work to police national identities in ways that go beyond being born in a place, which, unlike in parts of Europe, is sufficient for U.S. citizenship. Several years ago, at the urging of my daughter, my father-in-law, who was born and grew up in Berlin, tried to reclaim his German citizenship under repatriation laws. His request was denied because he had never actively claimed this citizenship during the Nazi period, when he was a teenager. He was told that he was Polish, the country of origin of his parents. But he would be unable to claim Polish citizenship because, for that, knowledge of Polish is required; Yiddish was not then, nor is it now, considered a proper language of this nation, which was just the problem in the first place, now compounded. (As long as monoculture laws stand in Poland, the stain of the destruction of the Jews will be ineradicable.)

    The power of American poetry comes from the mixing of many languages and the resistance to the dominance of any one language, including English – or anyway and one kind of English. As I argue in “The Poetics of the Americas” in My Way: Speeches and Poems, it is the overturning of standard English by second languages and vernacular/dialect speakers that defines American poetry, which is not to say that there is no resistance to this idea in America.

    The question of who owns a nation or a people is not, of course, just one of language. We can say in the U.S. that Occupy Wall Street raised the specter of the one percent whose control is through cultural tolerance combined with economic dominance. But the closer analogy in the U.S. for the systematic extermination process here is the mass incarceration of African-American young men, stripping them of opportunities to fully participate in American cultural and economic life.1 In Poland, you have about 224 prisoners per 100,000 people, in the US it’s over 750 but for some age groups of black men it’s over 10,000, one in nine people in that population. More than one in three young black men without a high school diploma is currently behind bars.2

    In Warsaw, we found an extraordinary new museum created as if from ashes. POLIN: The Museum of the History of Polish Jews is a site-specific work, situated on the ground of the Jewish Ghetto and as such serving as a memorial as well as a tribute to those who fought in the uprising. The museum’s main entrance is directly across from the more traditional and figurative “Monument to the Ghetto Heroes”, which was erected in 1948. The meaning of the museum is fundamentally connected to its location: as we descended into the core exhibition on the lower level, a ghostly presence makes itself felt and was a constant, and welcome, companion. This is a museum not of artifacts but of the historical record. POLIN overwhelms not with abstract splendor or holocaust memorial kitsch (as in the statue in front of Berlin’s Friedrichstrasse station, of children going off to “camp” as if the camp was “summer” and not “death”).3 This is a museum of deep, or thick, description. In place of sentimental, monumental, or abstract gestures of loss, icons that mark an absence, POLIN, in its core exhibition, curated by Barbara Kirchenbaum-Gimlet, provides a superfluity of what was present to substantiate (and transubstantiate) what has been lost. The museum is a not a funeral pyre but a living archive of what the Jews called Po-Lan when they arrived in the Middle Ages. And it succeeds in making indelible the presence of the Jews in and as Poland: not an eradicated blip in Polish history but as ineradicable thread in the fabric of Poland. Ineradicable even after being eradicated. As in Poe’s “Tell-tale Heart” –– under the floorboards that seem to seal us off from the ground, is a beating heart –– or here, let’s just say, once the floor board that covers the ground is removed, a vast underground cavern is revealed, as deep as the world and as wide as possibility. This cavern is not filled with broken rocks or statuettes of forlorn children. As we look into it more closely, we can see that it is a book.

    Jane Eisner’s “Chasing Ghosts, Reviving Spirits: The Fall and Rise of Poland’s Jews” provides account of a trip to Poland at the same time as ours; and, indeed, Eisner’s experiences were close to ours.4 A number of Polish commentators on this article lamented the author’s failure to blame the Nazi’s for the systematic extermination of the Polish Jews or that she failed to acknowledged the horrific suffering of many Polish people during the war. As Benjamin might have said, history is written by the (self-) righteous. He made me do it. POLIN brings the horse of denial to the water but it doesn’t make it drink. In that sense, the museum is programmatically gentle, a wise choice. There is much documentation of the role the Jewish ghetto police played in the extermination process. As for Polish complicity, there is a small placard that notes, with a bloodless tone, that Poles had their own battles to fight and their own everyday lives to live; most stood by and did nothing. (And that leaves out the active participation, and profit, on part of many Poles but also the active resistance on the part of some others.) Daddy, what did you do during the war? POLIN starkly documents the continuing anti-Semitic terror in Poland during what the Poles call the “Communist time”: under this new post-Nazi regime, most of the remaining Polish Jews left Poland – but not to the Poles but rather to what was left of them. The Jewish extermination in Poland was a Polish extermination or, let’s just say, the blood is mixed on the ground. It’s not a blame game, it’s a change game.

    I SURVIVED THE MUSEUM OF THE HISTORY OF POLISH JEWS

    On our final evening in Warsaw, we went to see the last remains of all the Jewish Ghetto, with bullet-ridden walls over which stenciled words say “never forget” and “truth.” There was one small peephole on a stone wall, similar to Duchamp’s peephole for “Étant donnés.” The view opened onto a dark, stagnant alley, caught in a time neither now nor then, a liminal space haunted by its own aching hollowness. On the other side of the alley was a window with a flickering light; someone was still there, barely there. In a year this last remnant of the ghetto will be covered over with new, chic shops, where perhaps you will be able to buy a figurine of an old Jew floating in amber. Maybe the one we nearly saw in the ghostly light that last night.

    _____

    notes:

    1. The House I Live In, a 2012 film by Eugene Jarecki, makes a compelling case for this analogy. Back to the essay

    2. economix.blogs.nytimes.com/2010/09/29/jail-and-jobs/?_php=true&_type=blogs&_r=0 Back to the essay

    3. One of the Kindertransport Sculptures by Frank Meilser, “Trains to Life – Trains to Death” (2008): “Standing at the Friedrichstrasse Railway Station this bronze sculpture commemorates 1.6 million children murdered in the Holocaust and 10,000 children whose lives were saved by being granted entry into England in 1938.” The site includes images and a line of Meisler’s Jewish kitsch figurines. Ten thousand divided by 1.6 million comes to .00625 (but whose counting?). Back to the essay

    4. Jewish Daily Forward, Nov. 30, 2014. forward.com/articles/209962/chasing-ghosts-reviving-spirits-the-fall-and-risee Back to the essay

    _____

  • Warding Off General Ludd: The Absurdity of “The Luddite Awards”

    Warding Off General Ludd: The Absurdity of “The Luddite Awards”

    By Zachary Loeb
    ~

    Of all the dangers looming over humanity no threat is greater than that posed by the Luddites.

    If the previous sentence seems absurdly hyperbolic, know that it only seems that way because it is, in fact, quite ludicrous. It has been over two hundred years since the historic Luddites rose up against “machinery hurtful to commonality,” but as their leader the myth enrobed General Ludd was never apprehended there are always those who fear that General Ludd is still out there, waiting with sledge hammer at the ready. True, there have been some activist attempts to revive the spirit of the Luddites (such as the neo-Luddites of the late 1980s and 1990s) – but in the midst of a society enthralled by (and in thrall to) smart phones, start-ups, and large tech companies – to see Luddites lurking in every shadow is a sign of either ideology, paranoia, or both.

    Yet, such an amusing mixture of unabashed pro-technology ideology and anxiety at the possibility of any criticism of technology is on full display in the inaugural “Luddite Awards” presented by The Information Technology and Innovation Foundation (ITIF). Whereas the historic Luddites needed sturdy hammers, and other such implements, to engage in machine breaking the ITIF seems to believe that the technology of today is much more fragile – it can be smashed into nothingness simply by criticism or even skepticism. As their name suggests, the ITIF is a think tank committed to the celebration of, and advocating for, technological innovation in its many forms. Thus it should not be surprising that a group committed to technological innovation would be wary of what it perceives as a growing chorus of “neo-Ludditism” that it imagines is planning to pull the plug on innovation. Therefore the ITIF has seen fit to present dishonorable “Luddite Awards” to groups it has deemed insufficiently enamored with innovation, these groups include (amongst others): The Vermont Legislature, The French Government, the organization Free Press, the National Rifle Association, and the Electronic Frontier Foundation. The ITIF “Luddite Awards” may mark the first time that any group has accused the Electronic Frontier Foundation of being a secret harbor for neo-Ludditism.

    luddite
    Unknown artist, “The Leader of the Luddites,” engraving, 1812 (image source: Wikipedia)

    The full report on “The 2014 ITIF Luddite Awards,” written by the ITIF’s president Robert D. Atkinson, presents the current state of technological innovation as being dangerously precarious. Though technological innovation is currently supplying people with all manner of devices, the ITIF warns against a growing movement born of neo-Ludditism that will aim to put a stop to further innovation. Today’s neo-Ludditism, in the estimation of the ITIF is distinct from the historic Luddites, and yet the goal of “ideological Ludditism” is still “to ‘smash’ today’s technology.” Granted, adherents of neo-Ludditism are not raiding factories with hammers, instead they are to be found teaching at universities, writing columns in major newspapers, disparaging technology in the media, and otherwise attempting to block the forward movement of progress. According to the ITIF (note the word “all”):

    “what is behind all ideological Ludditism is the general longing for a simpler life from the past—a life with fewer electronics, chemicals, molecules, machines, etc.” (ITIF, 3)

    Though the chorus of Ludditisim has, in the ITIF’s reckoning, grown to an unacceptable volume of late, the foundation is quick to emphasize that Ludditism is nothing new. What is new, as the ITIF puts it, is that these nefarious Luddite views have, apparently, moved from the margins and infected the larger public discourse around technology. A diverse array of figures and groups from figures like environmentalist Bill McKibben, conservative thinker James Pethokoukis, economist Paul Krugman, writers for Smithsonian Magazine, to foundations like Free Press, the EFF and the NRA – are all tarred with the epithet “Luddite.”The neo-Luddites, according to ITIF, issue warnings against unmitigated acceptance of innovation when they bring up environmental concerns, mention the possibility of jobs being displaced by technology, write somewhat approvingly of the historic Luddites, or advocate for Net Neutrality.

    While the ITIF holds to the popular, if historically inaccurate, definition of Luddite as “one who resists technological change,” their awards make clear that the ITIF would like to add to this definition the words “or even mildly opposes any technological innovation.” The ten groups awarded “Luddite Awards” are a mixture of non-profit public advocacy organizations and various governments – though the ITIF report seems to revel in attacking Bill McKibben he was not deemed worthy of an award (maybe next year). The awardees include: the NRA for opposing smart guns, The Vermont legislature for requiring the labeling of GMOS, Free Press’s support of net neutrality which is deemed as an affront to “smarter broadband networks,” news reports which “claim that ‘robots are killing jobs,” the EFF is cited as it “opposes Health IT,” and various governments in several states are reprimanded for “cracking down” on companies like Airbnb, Uber and Lyft. The ten recipients of Luddite awards may be quite surprised to find that they have been deemed adherents of neo-Ludditism, but in the view of the ITIF the actions these groups have taken indicate that General Ludd is slyly guiding their moves. Though the Luddite Awards may have a somewhat silly feeling, the ITIF cautions that the threat is serious, as the report ominously concludes:

    “But while we can’t stop the Luddites from engaging in their anti-progress, anti-innovation activities, we can recognize them for what they are: actions and ideas that are profoundly anti-progress, that if followed would mean a our children [sic] will live lives as adults nowhere near as good as the lives they could live if we instead embraced, rather than fought innovation.” (ITIF, 19)

    Credit is due to the ITIF for their ideological consistency. In putting together their list of recipients for the inaugural “Luddite Awards” – the foundation demonstrates that they are fully committed to technological innovation and they are unflagging in their support of that cause. Nevertheless, while the awards (and in particular the report accompanying the awards) may be internally ideologically consistent it is also a work of dubious historical scholarship, comical neoliberal paranoia, and evinces a profound anti-democratic tendency. Though the ITIF awards aim to target what it perceives as “neo-Ludditism” even a cursory glance at their awardees makes it abundantly clear that what the organization actually opposes is any attempt to regulate technology undertaken by a government, or advocated for by a public interest group. Even in a country as regulation averse as the contemporary United States it is still safer to defame Luddites than to simply state that you reject regulation. The ITIF carefully cloaks its ideology in the aura of terms with positive connotations such as “innovation,” “progress,” and “freedom” but these terms are only so much fresh paint over the same “free market” ideology that only values innovation, progress and freedom when they are in the service of neoliberal economic policies. Nowhere does the ITIF engage seriously with the questions of “who profits from this innovation?” “who benefits from this progress?” “is this ‘freedom’ equally distributed or does it reinforce existing inequities?” – the terms are used as ideological sledgehammers far blunter than any tool the Luddites ever used. This raw ideology is on perfect display in the very opening line of the award announcement, which reads:

    “Technological innovation is the wellspring of human progress, bringing higher standards of living, improved health, a cleaner environment, increased access to information and many other benefits.” (ITIF, 1)

    One can only applaud the ITIF for so clearly laying out their ideology at the outset, and one can only raise a skeptical eyebrow at this obvious case of the logical fallacy of Begging the Question. To claim that “technological innovation is the wellspring of human progress” is an assumption that demands proof, it is not a conclusion in and of itself. While arguments can certainly be made to support this assumption there is little in the ITIF report that suggests the ITIF is willing to engage in the type of critical reflection, which would be necessary for successfully supporting this argument (though, to be fair, the ITIF has published many other reports some of which may better lay out this claim). The further conclusions that such innovation brings “higher standards of living, improved health, a cleaner environment” and so forth are further assumptions that require proof – and in the process of demonstrating this proof one is forced (if engaging in honest argumentation) to recognize the validity of competing claims. Particularly as many of the “benefits” ITIF seeks to celebrate do not accrue evenly. True, an argument can be made that technological innovation has an important role to play in ushering in a “cleaner environment” – but tell that to somebody who lives next to an e-waste dump where mountains of the now obsolete detritus of “technological innovation” leach toxins into the soil. The ITIF report is filled with such pleasant sounding “common sense” technological assumptions that have been, at the very least, rendered highly problematic by serious works of inquiry and scholarship in the field of the history of technology. As classic works in the scholarly literature of the Science and Technology Studies field, such as Ruth Schwartz Cowan’s More Work for Mother, make clear “technological innovation” does not always live up to its claims. Granted, it is easy to imagine that the ITIF would offer a retort that simply dismisses all such scholarship as tainted by neo-Ludditism. Yet recognizing that not all “innovation” is a pure blessing does not represent a rejection of “innovation” as such – it just recognize that “innovation” is only one amongst many competing values a society must try to balance.

    Instead of engaging with critics of “technological innovation” in good faith, the ITIF jumps from one logical fallacy to another, trading circular reasoning for attacking the advocate. The author of the ITIF report seems to delight in pillorying Bill McKibben but also aims its barbs at scholars like David Noble and Neil Postman for exposing impressionable college aged minds to their “neo-Luddite” biases. That the ITIF seems unconcerned with business schools, start-up culture, and a “culture industry” that inculcates an adoration for “technological innovation” to the same “impressionable minds” is, obviously, not commented upon. However, if a foundation is attempting to argue that universities are currently a hotbed of “neo-Ludditism” than it is questionable why the ITIF should choose to signal out two professors for special invective who are both deceased – Postman died in 2003 and David Noble died in 2010.

    It almost seems as if the ITIF report cites serious humanistic critics of “technological innovation” as a way to make it seem as though it has actually wrestled with the thought of such individuals. After all, the ITIF report deigns to mention two of the most prominent thinkers in the theoretical legacy of the critique of technology, Lewis Mumford and Jacques Ellul, but it only mentions them in order to dismiss them out of hand. The irony, naturally, is that thinkers like Mumford and Ellul (to say nothing of Postman and Noble) would have not been surprised in the least by the ITIF report as their critiques of technology also included a recognition of the ways that the dominant forces in technological society (be it in the form of Ellul’s “Technique” or Mumford’s “megamachine”) depended upon the ideological fealty of those who saw their own best interests as aligning with that of the new technological regimes of power. Indeed, the ideological celebrants of technology have become a sort of new priesthood for the religion of technology, though as Mumford quipped in Art and Technics:

    “If you fall in love with a machine there is something wrong with your love-life. If you worship a machine there is something wrong with your religion.” (Art and Technics, 81)

    Trade out the word “machine” in the above quotation with “technological innovation” and it applies perfectly to the ITIF awards document. And yet, playful gibes aside, there are many more (many, many more) barbs that one can imagine Mumford directing at the ITIF. As Mumford wrote in The Pentagon of Power:

    “Consistently the agents of the megamachine act as if their only responsibility were to the power system itself. The interests and demands of the populations subjected to the megamachine are not only unheeded but deliberately flouted.” (The Pentagon of Power, 271)

    The ITIF “Luddite Awards” are a pure demonstration of this deliberate flouting of “the interests and demands of the populations” who find themselves always on the receiving end of “technological innovation.” For the ITIF report shows an almost startling disregard for the concerns of “everyday people” and though the ITIF is a proudly nonpartisan organization the report demonstrates a disturbingly anti-democratic tendency. That the group does not lean heavily toward Democrats or Republicans only demonstrates the degree to which both parties eat from the same neoliberal trough – routinely filled with fresh ideological slop by think tanks like ITIF. Groups that advocate in the interest of their supporters in the public sphere (such as Free Press, the EFF, and the NRA {yes, even them}) are treated as interlopers worthy of mockery for having the audacity to raise concerns; similarly elected governmental bodies are berated for daring to pass timid regulations. The shape of the “ideal society” that one detects in the ITIF report is one wherein “technological innovation” knows no limits, and encounters no opposition, even if these limits are relatively weak regulations or simply citizens daring to voice a contrary opinion – consequences be damned! On the high-speed societal train of “technological innovation” the ITIF confuses a few groups asking for a slight reduction of speed with groups threatening to derail the train.

    Thus the key problem of the ITIF “Luddite Awards” emerges – and it is not simply that the ITIF continues to use Luddite as an epithet – it is that the ITIF seems willfully ignorant of any ethical imperatives other than a broadly defined love of “technological innovation.” In handing out “Luddite Awards” the ITIF reveals that it recognizes “technological innovation” as the crowning example of “the good.” It is not simply one “good” amongst many that must carefully compromise with other values (such as privacy, environmental concerns, labor issues, and so forth), rather it is the definitive and ultimate case of “the good.” This is not to claim that “technological innovation” is not amongst values that represent “the good,” but it is not the only value – treating it as such lead to confusing (to borrow a formulation from Lewis Mumford) “the goods life with the good life.” By fully privileging “technological innovation” the ITIF treats other values and ethical claims as if they are to be discarded – the philosopher Hans Jonas’s The Imperative of Responsibility (which advocated for a cautious approach to technological innovation that emphasized the potential risks inherent in new technologies) is therefore tossed out the window to be replaced by “the imperative of innovation” along with a stack of business books and perhaps an Ayn Rand novel, or two, for good measure.

    Indeed, responsibility for the negative impacts of innovation is shrugged off in the ITIF awards, even as many of the awardees (such as the various governments) wrestle with the responsibility that tech companies seem to so happily flaunt. The disrupters hate being disrupted. Furthermore, as should come as no surprise, the ITIF report maintains an aura that smells strongly of colonialism and disregard for the difficulties faced by those who are “disrupted” by “technological innovation.” The ITIF may want to reprimand organizations for trying to gently slow (which is not the same as stopping) certain forms of “technological innovation,” but the report has nothing to say about those who work mining the coltan that powers so many innovative devices, no concern for the factory workers who assemble these devices, and – of course – nothing to say about e-waste. Evidently to think such things are worthy of concern, to even raise the issue of consequences, is a sign of Ludditism. The ITIF holds out the promise of “better days ahead” and shows no concern for those whose lives must be trampled upon in the process. Granted, it is easy to ignore such issues when you work for a think tank in Washington DC and not as a coltan miner, a device assembler, a resident near an e-waste dump, or an individual whose job has just been automated.

    The ITIF “Luddite Awards” are yet another installment of the tech world/business press game of “Who’s Afraid of General Ludd” in which the group shouting the word “Luddite” at all opponents reveals that it has a less nuanced understanding of technology than was had by the historic Luddites. After all, the Luddites were not opposed to technology as such, nor were they opposed to “technological innovation,” rather, as E.P. Thompson describes in The Making of the English Working Class:

    “What was at issue was the ‘freedom’ of the capitalist to destroy the customs of the trade, whether by new machinery, by the factory-system, or by unrestricted competition, beating-down wages, undercutting his rivals, and undermining standards of craftsmanship…They saw laissez faire, not as freedom but as ‘foul Imposition”. They could see no ‘natural law’ by which one man, or a few men, could engage in practices which brought manifest injury to their fellows.” (Thompson, 548)

    What is at issue in the “Luddite Awards” is the “freedom” of “technological innovators” (the same-old “capitalists”) to force their priorities upon everybody else – and while the ITIF may want to applaud such “freedom” it is clear that they do not intend to extend such freedom to the rest of the population. The fear that can be detected in the ITIF “Luddite Awards” is not ultimately directed at the award recipients, but at an aspect of the historic Luddites that the report seems keen on forgetting: namely, that the Luddites organized a mass movement that enjoyed incredible popular support – which was why it was ultimately the military (not “seeing the light” of “technological innovation”) that was required to bring the Luddite uprisings to a halt. While it is questionable whether many of the recipients of “Luddite Awards” will view the award as an honor, the term “Luddite” can only be seen as a fantastic compliment when it is used as a synonym for a person (or group) that dares to be concerned with ethical and democratic values other than a simple fanatical allegiance to “technological innovation.” Indeed, what the ITIF “Luddite Awards” demonstrate is the continuing veracity of the philosopher Günther Anders statement, in the second volume of The Obsolescence of Man, that:

    “In this situation, it is no use to brandish scornful words like ‘Luddites’. If there is anything that deserves scorn it is, to the contrary, today’s scornful use of the term, ‘Luddite’ since this scorn…is currently more obsolete than the allegedly obsolete Luddism.” (Anders, Introduction – Section 7)

    After all, as Anders might have reminded the people at ITIF: gas chambers, depleted uranium shells, and nuclear weapons are also “technological innovations.”

    Works Cited

    • Anders, Günther. The Obsolescence of Man: Volume II – On the Destruction of Life in the Epoch of the Third Industrial Revolution. (translated by Josep Monter Pérez, Pre-Textos, Valencia, 2011). Available online: here.
    • Atkinson, Robert D. The 2014 Luddite Awards. January 2015.
    • Mumford, Lewis. The Myth of the Machine, volume 2 – The Pentagon of Power. New York: Harvest/Harcourt Brace Jovanovich, 1970.
    • Mumford, Lewis. Art and Technics. New York: Columbia University Press, 2000.
    • Thompson, E.P. The Making of the English Working Class. New York: Vintage Books, 1966.
    • Not cited but worth a look – Eric Hobsbawm’s classic article “The Machine Breakers.”


    _____

    Zachary Loeb is a writer, activist, librarian, and terrible accordion player. He earned his MSIS from the University of Texas at Austin, and is currently working towards an MA in the Media, Culture, and Communications department at NYU. His research areas include media refusal and resistance to technology, ethical implications of technology, alternative forms of technology, and libraries as models of resistance. Using the moniker “The Luddbrarian,” Loeb writes at the blog LibrarianShipwreck, where this post first appeared. He is a frequent contributor to The b2 Review Digital Studies section.

    Back to the essay

  • The Reticular Fallacy

    The Reticular Fallacy

    By Alexander R. Galloway
    ~

    We live in an age of heterogenous anarchism. Contingency is king. Fluidity and flux win over solidity and stasis. Becoming has replaced being. Rhizomes are better than trees. To be political today, one must laud horizontality. Anti-essentialism and anti-foundationalism are the order of the day. Call it “vulgar ’68-ism.” The principles of social upheaval, so associated with the new social movements in and around 1968, have succeed in becoming the very bedrock of society at the new millennium.

    But there’s a flaw in this narrative, or at least a part of the story that strategically remains untold. The “reticular fallacy” can be broken down into two key assumptions. The first is an assumption about the nature of sovereignty and power. The second is an assumption about history and historical change. Consider them both in turn.

    (1) First, under the reticular fallacy, sovereignty and power are defined in terms of verticality, centralization, essence, foundation, or rigid creeds of whatever kind (viz. dogma, be it sacred or secular). Thus the sovereign is the one who is centralized, who stands at the top of a vertical order of command, who rests on an essentialist ideology in order to retain command, who asserts, dogmatically, unchangeable facts about his own essence and the essence of nature. This is the model of kings and queens, but also egos and individuals. It is what Barthes means by author in his influential essay “Death of the Author,” or Foucault in his “What is an Author?” This is the model of the Prince, so often invoked in political theory, or the Father invoked in psycho-analytic theory. In Derrida, the model appears as logos, that is, the special way or order of word, speech, and reason. Likewise, arkhe: a term that means both beginning and command. The arkhe is the thing that begins, and in so doing issues an order or command to guide whatever issues from such a beginning. Or as Rancière so succinctly put it in his Hatred of Democracy, the arkhe is both “commandment and commencement.” These are some of the many aspects of sovereignty and power as defined in terms of verticality, centralization, essence, and foundation.

    (2) The second assumption of the reticular fallacy is that, given the elimination of such dogmatic verticality, there will follow an elimination of sovereignty as such. In other words, if the aforementioned sovereign power should crumble or fall, for whatever reason, the very nature of command and organization will also vanish. Under this second assumption, the structure of sovereignty and the structure of organization become coterminous, superimposed in such a way that the shape of organization assumes the identical shape of sovereignty. Sovereign power is vertical, hence organization is vertical; sovereign power is centralized, hence organization is centralized; sovereign power is essentialist, hence organization, and so on. Here we see the claims of, let’s call it, “naïve” anarchism (the non-arkhe, or non foundation), which assumes that repressive force lies in the hands of the bosses, the rulers, or the hierarchy per se, and thus after the elimination of such hierarchy, life will revert so a more direct form of social interaction. (I say this not to smear anarchism in general, and will often wish to defend a form of anarcho-syndicalism.) At the same time, consider the case of bourgeois liberalism, which asserts the rule of law and constitutional right as a way to mitigate the excesses of both royal fiat and popular caprice.

    reticular connective tissue
    source: imgkid.com

    We name this the “reticular” fallacy because, during the late Twentieth Century and accelerating at the turn of the millennium with new media technologies, the chief agent driving the kind of historical change described in the above two assumptions was the network or rhizome, the structure of horizontal distribution described so well in Deleuze and Guattari. The change is evident in many different corners of society and culture. Consider mass media: the uni-directional broadcast media of the 1920s or ’30s gradually gave way to multi-directional distributed media of the 1990s. Or consider the mode of production, and the shift from a Fordist model rooted in massification, centralization, and standardization, to a post-Fordist model reliant more on horizontality, distribution, and heterogeneous customization. Consider even the changes in theories of the subject, shifting as they have from a more essentialist model of the integral ego, however fraught by the volatility of the unconscious, to an anti-essentialist model of the distributed subject, be it postmodernism’s “schizophrenic” subject or the kind of networked brain described by today’s most advanced medical researchers.

    Why is this a fallacy? What is wrong about the above scenario? The problem isn’t so much with the historical narrative. The problem lies in an unwillingness to derive an alternative form of sovereignty appropriate for the new rhizomatic societies. Opponents of the reticular fallacy claim, in other words, that horizontality, distributed networks, anti-essentialism, etc., have their own forms of organization and control, and indeed should be analyzed accordingly. In the past I’ve used the concept of “protocol” to describe such a scenario as it exists in digital media infrastructure. Others have used different concepts to describe it in different contexts. On the whole, though, opponents of the reticular fallacy have not effectively made their case, myself included. The notion that rhizomatic structures are corrosive of power and sovereignty is still the dominant narrative today, evident across both popular and academic discourses. From talk of the “Twitter revolution” during the Arab Spring, to the ideologies of “disruption” and “flexibility” common in corporate management speak, to the putative egalitarianism of blog-based journalism, to the growing popularity of the Deleuzian and Latourian schools in philosophy and theory: all of these reveal the contemporary assumption that networks are somehow different from sovereignty, organization, and control.

    To summarize, the reticular fallacy refers to the following argument: since power and organization are defined in terms of verticality, centralization, essence, and foundation, the elimination of such things will prompt a general mollification if not elimination of power and organization as such. Such an argument is false because it doesn’t take into account the fact that power and organization may inhabit any number of structural forms. Centralized verticality is only one form of organization. The distributed network is simply a different form of organization, one with its own special brand of management and control.

    Consider the kind of methods and concepts still popular in critical theory today: contingency, heterogeneity, anti-essentialism, anti-foundationalism, anarchism, chaos, plasticity, flux, fluidity, horizontality, flexibility. Such concepts are often praised and deployed in theories of the subject, analyses of society and culture, even descriptions of ontology and metaphysics. The reticular fallacy does not invalidate such concepts. But it does put them in question. We can not assume that such concepts are merely descriptive or neutrally empirical. Given the way in which horizontality, flexibility, and contingency are sewn into the mode of production, such “descriptive” claims are at best mirrors of the economic infrastructure and at worse ideologically suspect. At the same time, we can not simply assume that such concepts are, by nature, politically or ethically desirable in themselves. Rather, we ought to reverse the line of inquiry. The many qualities of rhizomatic systems should be understood not as the pure and innocent laws of a newer and more just society, but as the basic tendencies and conventional rules of protocological control.


    _____

    Alexander R. Galloway is a writer and computer programer working on issues in philosophy, technology, and theories of mediation. Professor of Media, Culture, and Communication at New York University, he is author of several books and dozens of articles on digital media and critical theory, including Protocol: How Control Exists after Decentralization (MIT, 2006), Gaming: Essays in Algorithmic Culture (University of Minnesota, 2006); The Interface Effect (Polity, 2012), and most recently Laruelle: Against the Digital (University of Minnesota, 2014), reviewed here earlier in 2014. Galloway has recently been writing brief notes on media and digital culture and theory at his blog, on which this post first appeared.

    Back to the essay

  • Teacher Wars and Teaching Machines

    Teacher Wars and Teaching Machines

    teacher warsa review of Dana Goldstein, The Teacher Wars: A History of America’s Most Embattled Profession (Doubleday, 2014)
    by Audrey Watters
    ~

    Teaching is, according to the subtitle of education journalist Dana Goldstein’s new book, “America’s Most Embattled Profession.” “No other profession,” she argues, ”operates under this level of political scrutiny, not even those, like policing or social work, that are also tasked with public welfare and are paid for with public funds.”

    That political scrutiny is not new. Goldstein’s book The Teacher Wars chronicles the history of teaching at (what has become) the K–12 level, from the early nineteenth century and “common schools” — that is, before before compulsory education and public school as we know it today — through the latest Obama Administration education policies. It’s an incredibly well-researched book that moves from the feminization of the teaching profession to the recent push for more data-driven teacher evaluation, observing how all along the way, teachers have been deemed ineffectual in some way or another — failing to fulfill whatever (political) goals the public education system has demanded be met, be those goals be economic, civic, or academic.

    As Goldstein describes it, public education is a labor issue; and it has been, it’s important to note, since well before the advent of teacher unions.

    The Teacher Wars and Teaching Machines

    To frame education this way — around teachers and by extension, around labor — has important implications for ed-tech. What happens if we examine the history of teaching alongside the history of teaching machines? As I’ve argued before, the history of public education in the US, particularly in the 20th century, is deeply intertwined with various education technologies – film, TV, radio, computers, the Internet – devices that are often promoted as improving access or as making an outmoded system more “modern.” But ed-tech is frequently touted too as “labor-saving” and as a corrective to teachers’ inadequacies and inefficiencies.

    It’s hardly surprising, in this light, that teachers have long looked with suspicion at new education technologies. With their profession constantly under attack, many teacher are worried no doubt that new tools are poised to replace them. Much is said to quiet these fears, with education reformers and technologists insisting again and again that replacing teachers with tech is not the intention.

    And yet the sentiment of science fiction writer Arthur C. Clarke probably does resonate with a lot of people, as a line from his 1980 Omni Magazine article on computer-assisted instruction is echoed by all sorts of pundits and politicians: “Any teacher who can be replaced by a machine should be.”

    Of course, you do find people like former Washington DC mayor Adrian Fenty – best known arguably via his school chancellor Michelle Rhee – who’ll come right out and say to a crowd of entrepreneurs and investors, “If we fire more teachers, we can use that money for more technology.”

    So it’s hard to ignore the role that technology increasingly plays in contemporary education (labor) policies – as Goldstein describes them, the weakening of teachers’ tenure protections alongside an expansion of standardized testing to measure “student learning,” all in the service finding and firing “bad teachers.” The growing data collection and analysis enabled by schools’ adoption of ed-tech feeds into the politics and practices of employee surveillance.

    Just as Goldstein discovered in the course of writing her book that the current “teacher wars” have a lengthy history, so too does ed-tech’s role in the fight.

    As Sidney Pressey, the man often credited with developing the first teaching machine, wrote in 1933 (from a period Goldstein links to “patriotic moral panics” and concerns about teachers’ political leanings),

    There must be an “industrial revolution” in education, in which educational science and the ingenuity of educational technology combine to modernize the grossly inefficient and clumsy procedures of conventional education. Work in the schools of the school will be marvelously though simply organized, so as to adjust almost automatically to individual differences and the characteristics of the learning process. There will be many labor-saving schemes and devices, and even machines — not at all for the mechanizing of education but for the freeing of teacher and pupil from the educational drudgery and incompetence.

    Or as B. F. Skinner, the man most associated with the development of teaching machines, wrote in 1953 (one year before the landmark Brown v Board of Education),

    Will machines replace teachers? On the contrary, they are capital equipment to be used by teachers to save time and labor. In assigning certain mechanizable functions to machines, the teacher emerges in his proper role as an indispensable human being. He may teach more students than heretofore — this is probably inevitable if the world-wide demand for education is to be satisfied — but he will do so in fewer hours and with fewer burdensome chores.

    These quotations highlight the longstanding hopes and fears about teaching labor and teaching machines; they hint too at some of the ways in which the work of Pressey and Skinner and others coincides with what Goldstein’s book describes: the ongoing concerns about teachers’ politics and competencies.

    The Drudgery of School

    One of the things that’s striking about Skinner and Pressey’s remarks on teaching machines, I think, is that they recognize the “drudgery” of much of teachers’ work. But rather than fundamentally change school – rather than ask why so much of the job of teaching entails “burdensome chores” – education technology seems more likely to offload that drudgery to machines. (One of the best contemporary examples of this perhaps: automated essay grading.)

    This has powerful implications for students, who – let’s be honest – suffer through this drudgery as well.

    Goldstein’s book doesn’t really address students’ experiences. Her history of public education is focused on teacher labor more than on student learning. As a result, student labor is missing from her analysis. This isn’t a criticism of the book; and it’s not just Goldstein that does this. Student labor in the history of public education remains largely under-theorized and certainly underrepresented. Cue AFT president Al Shanker’s famous statement: “Listen, I don’t represent children. I represent the teachers.”

    But this question of student labor seems to be incredibly important to consider, particularly with the growing adoption of education technologies. Students’ labor – students’ test results, students’ content, students’ data – feeds the measurements used to reward or punish teachers. Students’ labor feeds the algorithms – algorithms that further this larger narrative about teacher inadequacies, sure, and that serve to financially benefit technology, testing, and textbook companies, the makers of today’s “teaching machines.”

    Teaching Machines and the Future of Collective Action

    The promise of teaching machines has long been to allow students to move “at their own pace” through the curriculum. “Personalized learning,” it’s often called today (although the phrase often refers only to “personalization” in terms of the pace, not in terms of the topics of inquiry). This means, supposedly, that instead of whole class instruction, the “work” of teaching changes: in the words of one education reformer, “with the software taking up chores like grading math quizzes and flagging bad grammar, teachers are freed to do what they do best: guide, engage, and inspire.”

    Again, it’s not clear how this changes the work of students.

    So what are the implications – not just pedagogically but politically – of students, their headphones on staring at their individual computer screens working alone through various exercises? Because let’s remember: teaching machines and all education technologies are ideological. What are the implications – not just pedagogically but politically – of these technologies’ emphasis on individualism, self-management, personal responsibility, and autonomy?

    What happens to discussion and debate, for example, in a classroom of teaching machines and “personalized learning”? What happens, in a world of schools catered to individual student achievement, to the community development that schools (at their best, at least) are also tasked to support?

    What happens to organizing? What happens to collective action? And by collectivity here, let’s be clear, I don’t mean simply “what happens to teachers’ unions”? If we think about The Teacher Wars and teaching machines side-by-side, we should recognize our analysis of (our actions surrounding) the labor issues of school need to go much deeper and more farther than that.

    _____

    Audrey Watters is a writer who focuses on education technology – the relationship between politics, pedagogy, business, culture, and ed-tech. She has worked in the education field for over 15 years: teaching, researching, organizing, and project-managing. Although she was two chapters into her dissertation (on a topic completely unrelated to ed-tech), she decided to abandon academia, and she now happily fulfills the one job recommended to her by a junior high aptitude test: freelance writer. Her stories have appeared on NPR/KQED’s education technology blog MindShift, in the data section of O’Reilly Radar, on Inside Higher Ed, in The School Library Journal, in The Atlantic, on ReadWriteWeb, and Edutopia. She is the author of the recent book The Monsters of Education Technology (Smashwords, 2014) and working on a book called Teaching Machines. She maintains the widely-read Hack Education blog, on which an earlier version of this review first appeared.

    Back to the essay

  • "Internet Freedom": Digital Empire?

    "Internet Freedom": Digital Empire?

    Dan Schiller, Digital Depression: Information Technology and Economic Crisisa review of Dan Schiller, Digital Depression: Information Technology and Economic Crisis  (University of Illinois Press, 2014)
    by Richard Hill
    ~
    Disclosure: the author of this review is mentioned in the Acknowledgements section of the reviewed book.

     

     

     

     

     

    Computers and telecommunications have revolutionized and disrupted all aspects of human activity, and even behavior. The impacts are broad and profound, with important consequences for governments, businesses, non-profit activities, and individuals. Networks of interconnected computer systems are driving many disruptive changes in business practices, information flows, and financial flows. Foremost amongst those networks is the Internet, much of which is global, or at least trans-national.

    According to some, the current governance arrangement for the Internet is nearly ideal. In particular, its global multi-stakeholder model of governance has resulted in a free and open Internet, which has enabled innovation and driven economic growth and well-being around the world. Others are of the view that things have not worked out that well. In particular, the Internet has resulted in mass surveillance by governments and by private companies, in monopolization, commodification and monetization of information and knowledge, in inequitable flows of finances between poor and rich countries, and in erosion of cultural diversity; further, those with central positions of influence have used it to consolidate power and to establish a new global regime of control and exploitation, under the guise of favoring liberalization, while in reality reinforcing the dominance and profitability of major corporations at the expense of the public interest, and the overarching position of certain national interests at the expense of global interests and well being.  [1]

    Dan Schiller’s book helps us to understand how rational and well-informed people can hold such diametrically opposing views. Schiller dissects the history of the growth of recent telecommunications networks and shows how they have significantly (indeed, dramatically) affected economic and political power relations around the world. And how, at the same time, US policies have consistently favored capital over labor, and have resulted in transfers of vast sums from developing countries to developed countries (in particular through interest on loans).

    2013 Berlin PRISM Demonstrations
    Participants wearing Edward Snowden and Chelsea Manning masks at 2013 Berlin protests against NSA PRISM program (image source: Wikipedia)

    Schiller documents in some detail how US policies that ostensibly promote the free flow of information around the world, the right of all people to connect to the Internet, and free speech, are in reality policies that have, by design, furthered the geo-economic and geo-political goals of the US, including its military goals, its imperialist tendencies, and the interests of large private companies based (if not always headquartered, at least for tax purposes) in the US. For example, strict copyright protection is held to be consistent with the free flow of information, as is mass surveillance. Cookies and exploitation of users’ personal data by Internet companies are held to be consistent with privacy rights (indeed, as Schiller shows, the US essentially denies the existence of the right to personal privacy for anything related to the Internet). There should be no requirements that data be stored locally, lest it escape the jurisdiction of the US surveillance apparatus. And very high profits and dominant positions in key Internet markets do not spark anti-trust or competition law investigations, as they might in any other industry.

    As Schiller notes, great powers have historically used communication systems to further their economic and strategic interests, so why should the US not so use the Internet? Thus stated, the matter seems obvious. But the matter is rarely thus stated. On the contrary, the Internet is often touted as a generous gift to the world’s people, able to lift them out of poverty and oppression, and to bring them the benefits of democracy and (or) free markets. Schiller’s carefully researched analysis is thus an important contribution.

    Schiller provides context by tracing the origins of the current financial and economic crises, pointing out that it is paradoxical that growing investments in Information and Communication Technologies (ICTs), and the supposed resultant productivity gains, did not prevent a major global economic crisis. Schiller explains how transnational corporations demanded liberalization of the terms on which they could use their private networks, and received then, resulting in profound changes in commodity chains, that is, the flow of production of goods and services. In particular, there has been an increase in transnational production, and this has reinforced the importance of transnational corporations. Further, ICTs have changed the nature of labor’s contribution to production, enabling many tasks to be shifted to unskilled workers (or even to consumers themselves: automatic teller machines (ATMs), for example, turn each of us into a bank clerk). However, the growth of the Internet did not transcend the regular economy: on the contrary, it was wrapped into the economy’s crisis tendencies and even exacerbated them.

    Schiller gives detailed accounts of these transformations in the automotive and financial industries, and in the military. The study of the effect of ICTs on the military is of particular interest considering that the Internet was originally developed as a military project, and that it is currently used by US intelligence agencies as a prime medium for the collection of information.

    Schiller then turns to telecommunications, explaining the very significant changes that took place in the USA starting in the late 1970s. Those changes resulted in a major restructuring of the dominant telecommunications playing field in the US and ultimately led to the growth of the Internet, a development which had world-wide effects. Schiller carefully describes the various US government actions that initiated and nurtured those changes, and that were instrumental in exporting similar changes to the rest of the world.

    Next, he analyzes how those changes affected and enabled the production of the networks themselves, the hardware used to build the networks and to use them (e.g. smartphones), and the software and applications that we all use today.

    Moving further up the value chain, Schiller explains how data-mining, coupled with advertising, fuels the growth of the dominant Internet companies, and how this data-mining is made possible only by denying data privacy, and how states use the very same techniques to implement mass surveillance.

    Having described the situation, Schiller proceeds to analyze it from economic and political perspectives. Given that the US was an early adopter of the Internet, it is not surprising that, because of economies of scale and network effects, US companies dominate the field (except in China, as Schiller explains in detail). Schiller describes how, given the influence of US companies on US politics, US policies, both domestic and foreign, are geared to allowing, or in fact favoring, ever-increasing concentration in key Internet markets, which is to the advantage of the US and its private companies–and despite the easy cant about decentralization and democratization.

    The book describes how the US views the Internet as an extraterritorial domain, subject to no authority except that of the US government and that of the dominant US companies. Each dictates its own law in specific spheres (for example, the US government has supervised, up to now, the management of Internet domain names and addresses; while US companies dictate unilateral terms and conditions to their users, terms and conditions that imply that users give up essentially all rights to their private data).

    Schiller describes how this state of affairs has become a foreign policy objective, with the US being willing to incur significant criticism and to pay a significant political price in order to maintain the status quo. That status quo is referred to as “the multi-stakeholder model”, in which private companies are essentially given veto power over government decisions (or at least over the decisions of any government other than the US government), a system that can be referred to as “corporatism”. Not only does the US staunchly defend that model for the Internet, it even tries to export it to other fields of human activity. And this despite, or perhaps because, that system allows companies to make profits when possible (in particular by exploiting state-built infrastructure or guarantees), and to transfer losses to states when necessary (as for example happened with the banking crisis).

    Schiller carefully documents how code words such as “freedom of access” and “freedom of speech” are used to justify and promote policies that in fact merely serve the interests of major US companies and, at the same time, the interests of the US surveillance apparatus, which morphed from a cottage industry into a major component of the military-industrial complex thanks to the Internet. He shows how the supposed open participation in key bodies (such as the Internet Engineering Task Force) is actually a screen to mask the fact that decisions are heavily influenced by insiders affiliated with US companies and/or the US government, and by agencies bound to the US as a state.

    As Schiller explains, this increasing dominance of US business and US political imperialism have not gone unchallenged, even if the challenges to date have mostly been rhetorical (again, except for China). Conflicts over Internet governance are related to rivalries between competing geo-political and geo-economic blocks, rivalries which will likely increase if economic growth continues to be weak. The rivalries are both between nations and within nations, and some are only emerging right now (for example, how to tax the digital economy, or the apparent emerging divergence of views between key US companies and the US government regarding mass surveillance).

    Indeed, the book explains how the challenges to US dominance have become more serious in the wake of the Snowden revelations, which have resulted in a significant loss of market share for some of the key US players, in particular with respect to cloud computing services. Those losses may have begun to drive the tip of a wedge between the so-far congruent goals of US companies and the US government

    In a nutshell, one can sum up what Schiller describes by paraphrasing Marx: “Capitalists of the world, unite! You have nothing to lose but the chains of government regulation.” But, as Schiller hints in his closing chapter, the story is still unfolding, and just as things did not work out as Marx thought they would, so things may not work out as the forces that currently dominate the Internet wish they will. So the slogan for the future might well be “Internet users of the world, unite! You have nothing to lose but the chains of exploitation of your personal data.”

    This book, and its extensive references, will be a valuable reference work for all future research in this area. And surely there will be much future research, and many more historical analyses of what may well be some of the key turning points in the history of mankind: the transition from the industrial era to the information era and the disruptions induced by that transition.

    _____

    Richard Hill, an independent consultant based in Geneva, Switzerland, was formerly a senior official at the International Telecommunication Union (ITU). He has been involved in internet governance issues since the inception of the internet and is now an activist in that area, speaking, publishing, and contributing to discussions in various forums. Among other works he is the author of The New International Telecommunication Regulations and the Internet: A Commentary and Legislative History (Springer, 2014). An earlier version of this review first appeared on Newsclick.

    Back to the essay
    _____

    1. From item 11 of document WSIS+10/4/6 of the preparatory process for the WSIS+10 High Level Event, which provided “a special platform for high-ranking officials of WSIS (World Summit on the Information Society) stakeholders, government, private sector, civil society and international organizations to express their views on the achievements, challenges and recommendations on the implementation” of various earlier internet governance initiatives backed by the International Telecommunications Union (ITU), the United Nations specialized agency for information and communications technologies, and other participants in the global internet governance sphere.

    Back to the essay