• Michael Miller — Seeing Ourselves, Loving Our Captors: Mark Jarzombek’s Digital Stockholm Syndrome in the Post-Ontological Age

    Michael Miller — Seeing Ourselves, Loving Our Captors: Mark Jarzombek’s Digital Stockholm Syndrome in the Post-Ontological Age

    a review of Mark Jarzombek, Digital Stockholm Syndrome in the Post-Ontological Age (University of Minnesota Press Forerunners Series, 2016)

    by Michael Miller

    ~

    All existence is Beta, basically. A ceaseless codependent improvement unto death, but then death is not even the end. Nothing will be finalized. There is no end, no closure. The search will outlive us forever

    — Joshua Cohen, Book of Numbers

    Being a (in)human is to be a beta tester

    — Mark Jarzombek, Digital Stockholm Syndrome in the Post-Ontological Age

    Too many people have access to your state of mind

    — Renata Adler, Speedboat

    Whenever I read through Vilém Flusser’s vast body of work and encounter, in print no less, one of the core concepts of his thought—which is that “human communication is unnatural” (2002, 5)––I find it nearly impossible to shake the feeling that the late Czech-Brazilian thinker must have derived some kind of preternatural pleasure from insisting on the ironic gesture’s repetition. Flusser’s rather grim view that “there is no possible form of communication that can communicate concrete experience to others” (2016, 23) leads him to declare that the intersubjective dimension of communication implies inevitably the existence of a society which is, in his eyes, itself an unnatural institution. One can find all over in Flusser’s work traces of his life-long attempt to think through the full philosophical implications of European nihilism, and evidence of this intellectual engagement can be readily found in his theories of communication.

    One of Flusser’s key ideas that draws me in is his notion that human communication affords us the ability to “forget the meaningless context in which we are completely alone and incommunicado, that is, the world in which we are condemned to solitary confinement and death: the world of ‘nature’” (2002, 4). In order to help stave off the inexorable tide of nature’s muted nothingness, Flusser suggests that humans communicate by storing memories, externalized thoughts whose eventual transmission binds two or more people into a system of meaning. Only when an intersubjective system of communication like writing or speech is established between people does the purpose of our enduring commitment to communication become clear: we communicate in order “to become immortal within others (2016, 31). Flusser’s playful positing of the ironic paradox inherent in the improbability of communication—that communication is unnatural to the human but it is also “so incredibly rich despite its limitations” (26)––enacts its own impossibility. In a representatively ironic sense, Flusser’s point is that all we are able to fully understand is our inability to understand fully.

    As Flusser’s theory of communication can be viewed as his response to the twentieth-century’s shifting technical-medial milieu, his ideas about communication and technics eventually led him to conclude that “the original intention of producing the apparatus, namely, to serve the interests of freedom, has turned on itself…In a way, the terms human and apparatus are reversed, and human beings operate as a function of the apparatus. A man gives an apparatus instructions that the apparatus has instructed him to give” (2011,73).[1] Flusser’s skeptical perspective toward the alleged affordances of human mastery over technology is most assuredly not the view that Apple or Google would prefer you harbor (not-so-secretly). Any cursory glance at Wired or the technology blog at Insider Higher Ed, to pick two long-hanging examples, would yield a radically different perspective than the one Flusser puts forth in his work. In fact, Flusser writes, “objects meant to be media may obstruct communication” (2016, 45). If media objects like the technical apparatuses of today actually obstruct communication, then why are we so often led to believe that they facilitate it? And to shift registers just slightly, if everything is said to be an object of some kind—even technical apparatuses––then cannot one be permitted to claim daily communion with all kinds of objects? What happens when an object—and an object as obsolete as a book, no less—speaks to us? Will we still heed its call?

    ***

    Speaking in its expanded capacity as neither narrator nor focalized character, the book as literary object addresses us in a direct and antagonistic fashion in the opening line to Joshua Cohen’s 2015 novel Book of Numbers. “If you’re reading this on a screen, fuck off. I’ll only talk if I’m gripped with both hands” (5), the book-object warns. As Cohen’s narrative tells the story of a struggling writer named Joshua Cohen (whose backstory corresponds mostly to the historical-biographical author Joshua Cohen) who is contracted to ghostwrite the memoir of another Joshua Cohen (who is the CEO of a massive Google-type company named Tetration), the novel’s middle section provides an “unedited” transcript of the conversation between the two Cohens in which the CEO recounts his upbringing and tremendous business success in and around the Bay Area from the late 1970s up to 2013 of the narrative’s present. The novel’s Silicon Valley setting, nominal and characterological doubling, and structural narrative coupling of the two Cohens’ lives makes it all but impossible to distinguish the personal histories of Cohen-the-CEO and Cohen-the-narrator from the cultural history of the development of personal computing and networked information technologies. The history of one Joshua Cohen––or all Joshua Cohens––is indistinguishable from the history of intrusive computational/digital media. “I had access to stuff I shouldn’t have had access to, but then Principal shouldn’t have had such access to me—cameras, mics,” Cohen-the-narrator laments. In other words, as Cohen-the-narrator ghostwrites another Cohen’s memoir within the context of the broad history of personal computing and the emergence of algorithmic governance and surveillance, the novel invites us to consider how the history of an individual––or every individual, it does not really matter––is also nothing more or anything less than the surveilled history of its data usage, which is always written by someone or something else, the ever-present Not-Me (who just might have the same name as me). The Self is nothing but a networked repository of information to be mined in the future.

    While the novel’s opening line addresses its hypothetical reader directly, its relatively benign warning fixes reader and text in a relation of rancor. The object speaks![2] And yet tech-savvy twenty-first century readers are not the only ones who seem to be fed up with books; books too are fed up with us, and perhaps rightly so. In an age when objects are said to speak vibrantly and withdraw infinitely; processes like human cognition are considered to be operative in complex technical-computational systems; and when the only excuse to preserve the category of “subjective experience” we are able to muster is that it affords us the ability “to grasp how networks technically distribute and disperse agency,” it would seem at first glance that the second-person addressee of the novel’s opening line would intuitively have to be a reading, thinking subject.[3] Yet this is the very same reading subject who has been urged by Cohen’s novel to politely “fuck off” if he or she has chosen to read the text on a screen. And though the text does not completely dismiss its readers who still prefer “paper of pulp, covers of board and cloth” (5), a slight change of preposition in its title points exactly to what the book fears most of all: Book as Numbers. The book-object speaks, but only to offer an ominous admonition: neither the book nor its readers ought to be reducible to computable numbers.

    The transduction of literary language into digital bits eliminates the need for a phenomenological, reading subject, and it suggests too that literature––or even just language in a general sense––and humans in particular are ontologically reducible to data objects that can be “read” and subsequently “interpreted” by computational algorithms. As Cohen’s novel collapses the distinction between author, narrator, character, and medium, its narrator observes that “the only record of my one life would be this record of another’s” (9). But in this instance, the record of one’s (or another’s) life is merely the history of how personal computational technologies have effaced the phenomenological subject. How have we arrived at the theoretically permissible premise that “People matter, but they don’t occupy a privileged subject position distinct from everything else in the world” (Huehls 20)? How might the “turn toward ontology” in theory/philosophy be viewed as contributing to our present condition?

    * **

    Mark Jarzombek’s Digital Stockholm Syndrome in the Post-Ontological Age (2016) provides a brief, yet stylistically ironic and incisive interrogation into how recent iterations of post- or inhumanist theory have found a strange bedfellow in the rhetorical boosterism that accompanies the alleged affordances of digital technologies and big data. Despite the differences between these two seemingly unrelated discourses, they both share a particularly critical or diminished conception of the anthro- in “anthropocentrism” that borrows liberally from the postulates of the “ontological turn” in theory/philosophy (Rosenberg n.p.). While the parallels between these discourses are not made explicit in Jarzombek’s book, Digital Stockholm Syndrome asks us to consider how a shared commitment to an ontologically diminished view of “the human” that galvanizes both technological determinism’s anti-humanism and post- or inhumanist theory has found its common expression in recent philosophies of ontology. In other words, the problem Digital Stockholm Syndrome takes up is this: what kind of theory of ontology, Being, and to a lesser extent, subjectivity, appeals equally to contemporary philosophers and Silicon Valley tech-gurus? Jarzombek gestures toward such an inquiry early on: “What is this new ontology?” he asks, and “What were the historical situations that produced it? And how do we adjust to the realities of the new Self?” (x).

    A curious set of related philosophical commitments united by their efforts to “de-center” and occasionally even eject “anthropocentrism” from the critical conversation constitute some of the realities swirling around Jarzombek’s “new Self.”[4] Digital Stockholm Syndrome provocatively locates the conceptual legibility of these philosophical realities squarely within an explicitly algorithmic-computational historical milieu. By inviting such a comparison, Jarzombek’s book encourages us to contemplate how contemporary ontological thought might mediate our understanding of the historical and philosophical parallels that bind the tradition of in humanist philosophical thinking and the rhetoric of twenty-first century digital media.[5]

    In much the same way that Alexander Galloway has argued for a conceptual confluence that exceeds the contingencies of coincidence between “the structure of ontological systems and the structure of the most highly evolved technologies of post-Fordist capitalism” (347), Digital Stockholm Syndrome argues similarly that today’s world is “designed from the micro/molecular level to fuse the algorithmic with the ontological” (italics in original, x).[6] We now understand Being as the informatic/algorithmic byproduct of what ubiquitous computational technologies have gathered and subsequently fed back to us. Our personal histories––or simply the records of our data use (and its subsequent use of us)––comprise what Jarzombek calls our “ontic exhaust…or what data experts call our data exhaust…[which] is meticulously scrutinized, packaged, formatted, processed, sold, and resold to come back to us in the form of entertainment, social media, apps, health insurance, clickbait, data contracts, and the like” (x).

    The empty second-person pronoun is placed on equal ontological footing with, and perhaps even defined by, its credit score, medical records, 4G data usage, Facebook likes, and threefold of its Tweets. “The purpose of these ‘devices,’” Jarzombek writes, “is to produce, magnify, and expose our ontic exhaust” (25). We give our ontic exhaust away for free every time we log into Facebook because it, in return, feeds back to us the only sense of “self” we are able to identify as “me.”[7] If “who we are cannot be traced from the human side of the equation, much less than the analytic side.‘I’ am untraceable” (31), then why do techno-determinists and contemporary oracles of ontology operate otherwise? What accounts for their shared commitment to formalizing ontology? Why must the Self be tracked and accounted for like a map or a ledger?

    As this “new Self,” which Jarzombek calls the “Being-Global” (2), travels around the world and checks its bank statement in Paris or tags a photo of a Facebook friend in Berlin while sitting in a cafe in Amsterdam, it leaks ontic exhaust everywhere it goes. While the hoovering up of ontic exhaust by GPS and commercial satellites “make[s] us global,” it also inadvertently redefines Being as a question of “positioning/depositioning” (1). For Jarzombek, the question of today’s ontology is not so much a matter of asking “what exists?” but of asking “where is it and how can it be found?” Instead of the human who attempts to locate and understand Being, now Being finds us, but only as long as we allow ourselves to be located.

    Today’s ontological thinking, Jarzombek points out, is not really interested in asking questions about Being––it is too “anthropocentric.”[8] Ontology in the twenty-first century attempts to locate Being by gathering data, keeping track, tracking changes, taking inventory, making lists, listing litanies, crunching the numbers, and searching the database. “Can I search for it on Google?” is now the most important question for ontological thought in the twenty-first century.

    Ontological thinking––which today means ontological accounting, or finding ways to account for the ontologically actuarial––is today’s philosophical equivalent to a best practices for data management, except there is no difference between one’s data and one’s Self. Nonetheless, any ontological difference that might have once stubbornly separated you from data about you no longer applies. Digital Stockholm Syndrome identifies this shift with the formulation: “From ontology to trackology” (71).[9] The philosophical shift that has allowed data about the Self to become the ontological equivalent to the Self emerges out of what Jarzombek calls an “animated ontology.”

    In this “animated ontology,” subject position and object position are indistinguishable…The entire system of humanity is microprocessed through the grid of sequestered empiricism” (31, 29). Jarzombek is careful to distinguish his “animated ontology” from the recently rebooted romanticisms which merely turn their objects into vibrant subjects. He notes that “the irony is that whereas the subject (the ‘I’) remains relatively stable in its ability to self-affirm (the lingering by-product of the psychologizing of the modern Self), objectivity (as in the social sciences) collapses into the illusions produced by the global cyclone of the informatic industry” (28).”[10] By devising tricky new ways to flatten ontology (all of which are made via po-faced linguistic fiat), “the human and its (dis/re-)embodied computational signifiers are on equal footing”(32). I do not define my data, but my data define me.

    ***

    Digital Stockholm Syndrome asserts that what exists in today’s ontological systems––systems both philosophical and computational––is what can be tracked and stored as data. Jarzombek sums up our situation with another pithy formulation: “algorithmic modeling + global positioning + human scaling +computational speed=data geopolitics” (12). While the universalization of tracking technologies defines the “global” in Jarzombek’s Being-Global, it also provides us with another way to understand the humanities’ enthusiasm for GIS and other digital mapping platforms as institutional-disciplinary expressions of a “bio-chemo-techno-spiritual-corporate environment that feeds the Human its sense-of-Self” (5).

    Mark Jarzombek, Digital Stockholm Syndrome in the Post-Ontological Age

    One wonders if the incessant cultural and political reminders regarding the humanities’ waning relevance have moved humanists to reconsider the very basic intellectual terms of their broad disciplinary pursuits. How come it is humanities scholars who are in some cases most visibly leading the charge to overturn many decades of humanist thought? Has the internalization of this depleted conception of the human reshaped the basic premises of humanities scholarship, Digital Stockholm Syndrome wonders? What would it even mean to pursue a “humanities” purged of “the human?” And is it fair to wonder if this impoverished image of humanity has trickled down into the formation of new (sub)disciplines?”[11]

    In a late chapter titled “Onto-Paranoia,” Jarzombek finally arrives at a working definition of Digital Stockholm Syndrome: data visualization. For Jarzombek, data-visualization “has been devised by the architects of the digital world” to ease the existential torture—or “onto-torture”—that is produced by Security Threats (59). Security threats are threatening because they remind us that “security is there to obscure the fact that whole purpose is to produce insecurity” (59). When a system fails, or when a problem occurs, we need to be conscious of the fact that the system has not really failed; “it means that the system is working” (61).[12] The Social, the Other, the Not-Me—these are all variations of the same security threat, which is just another way of defining “indeterminacy” (66). So if everything is working the way it should, we rarely consider the full implications of indeterminacy—both technical and philosophical—because to do so might make us paranoid, or worse: we would have to recognize ourselves as (in)human subjects.

    Data-visualizations, however, provide a soothing salve which we can (self-)apply in order to ease the pain of our “onto-torture.” Visualizing data and creating maps of our data use provide us with a useful and also pleasurable tool with which we locate ourselves in the era of “post-ontology.”[13] “We experiment with and develop data visualization and collection tools that allow us to highlight urban phenomena. Our methods borrow from the traditions of science and design by using spatial analytics to expose patterns and communicating those results, through design, to new audiences,” we are told by one data-visualization project (http://civicdatadesignlab.org/).  As we affirm our existence every time we travel around the globe and self-map our location, we silently make our geo-data available for those who care to sift through it and turn it into art or profit.

    “It is a paradox that our self-aestheticizing performance as subjects…feeds into our ever more precise (self-)identification as knowable and predictable (in)human-digital objects,” Jarzombek writes. Yet we ought not to spend too much time contemplating the historical and philosophical complexities that have helped create this paradoxical situation. Perhaps it is best we do not reach the conclusion that mapping the Self as an object on digital platforms increases the creeping unease that arises from the realization that we are mappable, hackable, predictable, digital objects––that our data are us. We could, instead, celebrate how our data (which we are and which is us) is helping to change the world. “’Big data’ will not change the world unless it is collected and synthesized into tools that have a public benefit,” the same data visualization project announces on its website’s homepage.

    While it is true that I may be a little paranoid, I have finally rested easy after having read Digital Stockholm Syndrome because I now know that my data/I are going to good use.[14] Like me, maybe you find comfort in knowing that your existence is nothing more than a few pixels in someone else’s data visualization.

    _____

    Michael Miller is a doctoral candidate in the Department of English at Rice University. His work has appeared or is forthcoming in symplokē and the Journal of Film and Video.

    Back to the essay

    _____

    Notes

    [1] I am reminded of a similar argument advanced by Tung-Hui Hu in his A Prehistory of the Cloud (2016). Encapsulating Flusser’s spirit of healthy skepticism toward technical apparatuses, the situation that both Flusser and Hu fear is one in which “the technology has produced the means of its own interpretation” (xixx).

    [2] It is not my aim to wade explicitly into discussions regarding “object-oriented ontology” or other related philosophical developments. For the purposes of this essay, however, Andrew Cole’s critique of OOO as a “new occasionalism” will be useful. “’New occasionalism,’” Cole writes, “is the idea that when we speak of things, we put them into contact with one another and ourselves” (112). In other words, the speaking of objects makes them objectively real, though this is only possible when everything is considered to be an object. The question, though, is not about what is or is not an object, but is rather what it means to be. For related arguments regarding the relation between OOO/speculative realism/new materialism and mysticism, see Sheldon (2016), Altieri (2016), Wolfendale (2014), O’Gorman (2013), and to a lesser extent Colebrook (2013).

    [3] For the full set of references here, see Bennett (2010), Hayles (2014 and 2016), and Hansen (2015).

    [4] While I cede that no thinker of “post-humanism” worth her philosophical salt would admit the possibility or even desirability of purging the sins of “correlationism” from critical thought all together, I cannot help but view such occasional posturing with a skeptical eye. For example, I find convincing Barbara Herrnstein-Smith’s recent essay “Scientizing the Humanities: Shifts, Negotiations, Collisions,” in which she compares the drive in contemporary critical theory to displace “the human” from humanistic inquiry to the impossible and equally incomprehensible task of overcoming the “‘astro’-centrism of astronomy or the biocentrism of biology” (359).

    [5] In “Modest Proposal for the Inhuman,” Julian Murphet identifies four interrelated strands of post- or inhumanist thought that combine a kind of metaphysical speculation with a full-blown demolition of traditional ontology’s conceptual foundations. They are: “(1) cosmic nihilism, (2) molecular bio-plasticity, (3) technical accelerationism, and (4) animality. These sometimes overlapping trends are severally engaged in the mortification of humankind’s stubborn pretensions to mastery over the domain of the intelligible and the knowable in an era of sentient machines, routine genetic modification, looming ecological disaster, and irrefutable evidence that we share 99 percent of our biological information with chimpanzees” (653).

    [6] The full quotation from Galloway’s essay reads: “Why, within the current renaissance of research in continental philosophy, is there a coincidence between the structure of ontological systems and the structure of the most highly evolved technologies of post-Fordist capitalism? [….] Why, in short, is there a coincidence between today’s ontologies and the software of big business?” (347). Digital Stockholm Syndrome begins by accepting Galloway’s provocations as descriptive instead of speculative. We do not necessarily wonder in 2017 if “there is a coincidence between today’s ontologies and the software of big business”; we now wonder instead how such a confluence came to be.

    [7] Wendy Hui Kyun Chun makes a similar point in her 2016 monograph Updating to Remain the Same: Habitual New Media. She writes, “If users now ‘curate’ their lives, it is because their bodies have become archives” (x-xi). While there is not ample space here to discuss the  full theoretical implications of her book, Chun’s discussion of the inherently gendered dimension to confession, self-curation as self-exposition, and online privacy as something that only the unexposed deserve (hence the need for preemptive confession and self-exposition on the internet) in digital/social media networks is tremendously relevant to Jarzombek’s Digital Stockholm Syndrome, as both texts consider the Self as a set of mutable and “marketable/governable/hackable categories” (Jarzombek 26) that are collected without our knowledge and subsequently fed back to the data/media user in the form of its own packaged and unique identity. For recent similar variations of this argument, see Simanowski (2017) and McNeill (2012).

    I also think Chun’s book offers a helpful tool for thinking through recent confessional memoirs or instances of “auto-theory” (fictionalized or not) like Maggie Nelson’s The Argonauts (2015), Sheila Heti’s How Should a Person Be (2010), Marie Calloway’s what purpose did i serve in your life (2013), and perhaps to a lesser degree Tao Lin’s Richard Yates (2010), Taipei (2013), Natasha Stagg’s Surveys, and Ben Lerner’s Leaving the Atocha Station (2011) and 10:04 (2014). The extent to which these texts’ varied formal-aesthetic techniques can be said to be motivated by political aims is very much up for debate, but nonetheless, I think it is fair to say that many of them revel in the reveal. That is to say, via confession or self-exposition, many of these novels enact the allegedly performative subversion of political power by documenting their protagonists’ and/or narrators’ certain social/political acts of transgression. Chun notes, however, that this strategy of self-revealing performs “resistance as a form of showing off and scandalizing, which thrives off moral outrage. This resistance also mimics power by out-spying, monitoring, watching, and bringing to light, that is, doxing” (151). The term “autotheory,” which was has been applied to Nelson’s The Argonauts in particular, takes on a very different meaning in this context. “Autotheory” can be considered as a theory of the self, or a self-theorization, or perhaps even the idea that personal experience is itself a kind of theory might apply here, too. I wonder, though, how its meaning would change if the prefix “auto” was understood within a media-theoretical framework not as “self” but as “automation.” “Autotheory” becomes, then, an automatization of theory or theoretical thinking, but also a theoretical automatization; or more to the point: what if “autotheory” describes instead a theorization of the Self or experience wherein “the self” is only legible as the product of automated computational-algorithmic processes?

    [8] Echoing the critiques of “correlationism” or “anthropocentrism” or what have you, Jarzombek declares that “The age of anthrocentrism is over” (32).

    [9] Whatever notion of (self)identity the Self might find to be most palatable today, Jarzombek argues, is inevitably mediated via global satellites. “The intermediaries are the satellites hovering above the planet. They are what make us global–what make me global” (1), and as such, they represent the “civilianization” of military technologies (4). What I am trying to suggest is that the concepts and categories of self-identity we work with today are derived from the informatic feedback we receive from long-standing military technologies.

    [10] Here Jarzombek seems to be suggesting that the “object” in the “objectivity” of “the social sciences” has been carelessly conflated with the “object” in “object-oriented” philosophy. The prioritization of all things “objective” in both philosophy and science has inadvertently produced this semantic and conceptual slippage. Data objects about the Self exist, and thus by existing, they determine what is objective about the Self. In this new formulation, what is objective about the Self or subject, in other words, is what can be verified as information about the self. In Indexing It All: The Subject in the Age of Documentation, Information, and Data (2014), Ronald Day argues that these global tracking technologies supplant traditional ontology’s “ideas or concepts of our human manner of being” and have in the process “subsume[d] and subvert[ed] the former roles of personal judgment and critique in personal and social beings and politics” (1). While such technologies might be said to obliterate “traditional” notions of subjectivity, judgment, and critique, Day demonstrates how this simultaneous feeding-forward and feeding back of data-about-the-Self represents the return of autoaffection, though in his formulation self-presence is defined as information or data-about-the-self whose authenticity is produced when it is fact-checked against a biographical database (3)—self-presence is a presencing of data-about-the-Self. This is all to say that the Self’s informational “aboutness”–its representation in and as data–comes to stand in for the Self’s identity, which can only be comprehended as “authentic” in its limited metaphysical capacity as a general informatic or documented “aboutness.”

    [11] Flusser is again instructive on this point, albeit in his own idiosyncratic way­­. Drawing attention to the strange unnatural plurality in the term “humanities,” he writes, “The American term humanities appropriately describes the essence of these disciplines. It underscores that the human being is an unnatural animal” (2002, 3). The plurality of “humanities,” as opposed to the singular “humanity,” constitutes for Flusser a disciplinary admission that not only the category of “the human” is unnatural, but that the study of such an unnatural thing is itself unnatural, as well. I think it is also worth pointing out that in the context of Flusser’s observation, we might begin to situate the rise of “the supplemental humanities” as an attempt to redefine the value of a humanities education. The spatial humanities, the energy humanities, medical humanities, the digital humanities, etc.—it is not difficult to see how these disciplinary off-shoots consider themselves as supplements to whatever it is they think “the humanities” are up to; regardless, their institutional injection into traditional humanistic discourse will undoubtedly improve both(sub)disciplines, with the tacit acknowledgment being that the latter has just a little more to gain from the former in terms of skills, technical know-how, and data management. Many thanks to Aaron Jaffe for bringing this point to my attention.

    [12] In his essay “Algorithmic Catastrophe—The Revenge of Contingency,” Yuk Hui notes that “the anticipation of catastrophe becomes a design principle” (125). Drawing from the work of Bernard Stiegler, Hui shows how the pharmacological dimension of “technics, which aims to overcome contingency, also generates accidents” (127). And so “as the anticipation of catastrophe becomes a design principle…it no longer plays the role it did with the laws of nature” (132). Simply put, by placing algorithmic catastrophe on par with a failure of reason qua the operations of mathematics, Hui demonstrates how “algorithms are open to contingency” only insofar as “contingency is equivalent to a causality, which can be logically and technically deduced” (136). To take Jarzombek’s example of the failing computer or what have you, while the blue screen of death might be understood to represent the faithful execution of its programmed commands, we should also keep in mind that the obverse of Jarzombek’s scenario would force us to come to grips with how the philosophical implications of the “shit happens” logic that underpins contingency-as-(absent) causality “accompanies and normalizes speculative aesthetics” (139).

    [13] I am reminded here of one of the six theses from the manifesto “What would a floating sheep map?,” jointly written by the Floating Sheep Collective, which is a cohort of geography professors. The fifth thesis reads: “Map or be mapped. But not everything can (or should) be mapped.” The Floating Sheep Collective raises in this section crucially important questions regarding ownership of data with regard to marginalized communities. Because it is not always clear when to map and when not to map, they decide that “with mapping squarely at the center of power struggles, perhaps it’s better that not everything be mapped.” If mapping technologies operate as ontological radars–the Self’s data points help point the Self towards its own ontological location in and as data—then it is fair to say that such operations are only philosophically coherent when they are understood to be framed within the parameters outlined by recent iterations of ontological thinking and its concomitant theoretical deflation of the rich conceptual make-up that constitutes the “the human.” You can map the human’s data points, but only insofar as you buy into the idea that points of data map the human. See http://manifesto.floatingsheep.org/.

    [14]Mind/paranoia: they are the same word!”(Jarzombek 71).

    _____

    Works Cited

    • Adler, Renata. Speedboat. New York Review of Books Press, 1976.
    • Altieri, Charles. “Are We Being Materialist Yet?” symplokē 24.1-2 (2016):241-57.
    • Calloway, Marie. what purpose did i serve in your life. Tyrant Books, 2013.
    • Chun, Wendy Hui Kyun. Updating to Remain the Same: Habitual New Media. The MIT Press, 2016.
    • Cohen, Joshua. Book of Numbers. Random House, 2015.
    • Cole, Andrew. “The Call of Things: A Critique of Object-Oriented Ontologies.” minnesota review 80 (2013): 106-118.
    • Colebrook, Claire. “Hypo-Hyper-Hapto-Neuro-Mysticism.” Parrhesia 18 (2013).
    • Day, Ronald. Indexing It All: The Subject in the Age of Documentation, Information, and Data. The MIT Press, 2014.
    • Floating Sheep Collective. “What would a floating sheep map?” http://manifesto.floatingsheep.org/.
    • Flusser, Vilém. Into the Universe of Technical Images. Translated by Nancy Ann Roth. University of Minnesota Press, 2011.
    • –––. The Surprising Phenomenon of Human Communication. 1975. Metaflux, 2016.
    • –––. Writings, edited by Andreas Ströhl. Translated by Erik Eisel. University of Minnesota Press, 2002.
    • Galloway, Alexander R. “The Poverty of Philosophy: Realism and Post-Fordism.” Critical Inquiry 39.2 (2013): 347-366.
    • Hansen, Mark B.N. Feed Forward: On the Future of Twenty-First Century Media. Duke University Press, 2015.
    • Hayles, N. Katherine. “Cognition Everywhere: The Rise of the Cognitive Nonconscious and the Costs of Consciousness.” New Literary History 45.2 (2014):199-220.
    • –––. “The Cognitive Nonconscious: Enlarging the Mind of the Humanities.” Critical Inquiry 42 (Summer 2016): 783-808.
    • Herrnstein-Smith, Barbara. “Scientizing the Humanities: Shifts, Collisions, Negotiations.” Common Knowledge  22.3 (2016):353-72.
    • Heti, Sheila. How Should a Person Be? Picador, 2010.
    • Hu, Tung-Hui. A Prehistory of the Cloud. The MIT Press, 2016.
    • Huehls, Mitchum. After Critique: Twenty-First Century Fiction in a Neoliberal Age. Oxford University Press, 2016.
    • Hui, Yuk. “Algorithmic Catastrophe–The Revenge of Contingency.” Parrhesia 23(2015): 122-43.
    • Jarzombek, Mark. Digital Stockholm Syndrome in the Post-Ontological Age. University of Minnesota Press, 2016.
    • Lin, Tao. Richard Yates. Melville House, 2010.
    • –––. Taipei. Vintage, 2013.
    • McNeill, Laurie. “There Is No ‘I’ in Network: Social Networking Sites and Posthuman Auto/Biography.” Biography 35.1 (2012): 65-82.
    • Murphet, Julian. “A Modest Proposal for the Inhuman.” Modernism/Modernity 23.3(2016): 651-70.
    • Nelson, Maggie. The Argonauts. Graywolf P, 2015.
    • O’Gorman, Marcel. “Speculative Realism in Chains: A Love Story.” Angelaki: Journal of the Theoretical Humanities 18.1 (2013): 31-43.
    • Rosenberg, Jordana. “The Molecularization of Sexuality: On Some Primitivisms of the Present.” Theory and Event 17.2 (2014):  n.p.
    • Sheldon, Rebekah. “Dark Correlationism: Mysticism, Magic, and the New Realisms.” symplokē 24.1-2 (2016): 137-53.
    • Simanowski, Roberto. “Instant Selves: Algorithmic Autobiographies on Social Network Sites.” New German Critique 44.1 (2017): 205-216.
    • Stagg, Natasha. Surveys. Semiotext(e), 2016.
    • Wolfendale, Peter. Object Oriented Philosophy: The Noumenon’s New Clothes. Urbanomic, 2014.
  • Eugene Thacker – Weird, Eerie, and Monstrous: A Review of “The Weird and the Eerie” by Mark Fisher

    Eugene Thacker – Weird, Eerie, and Monstrous: A Review of “The Weird and the Eerie” by Mark Fisher

    by Eugene Thacker

    Review of Mark Fisher, The Weird and the Eerie (Repeater, 2017)

    For a long time, the horror genre was not generally considered worthy of critical, let alone philosophical, reflection; it was the stuff of cheap thrills, pulp magazines, B-movies. Much of this has changed in the ensuing years, as a robust and diverse critical literature has emerged around the horror genre, much of which not only considers the horror genre as a reflection of society, but as an autonomous platform for posing far-reaching questions concerning the fate of the humans species, the species that has named itself. These are sentiments that have preoccupied recent writing on the horror genre, much of which borrows from developments in contemporary philosophy, and is attempting to expand the confines of horror beyond the usual fixation on gore, violence, and shock tactics. This hasn’t always been the case. Even today, writing on genre horror often tends towards “list” books (of the type The Top 100 Italian Horror Films From 1977, Volume IV), or books that are basically print-on-demand databases (The Encyclopedia of Asian Ghost Stories from the Beginning of Time, and Before That). These are rounded out by a plethora of introductory textbooks and surveys, usually aimed at film studies undergraduates (e.g. Key Terms in Cultural Studies: Splatterpunk), and opaque academic monographs of Lacanian psychoanalytic semiotic readings of horror film that themselves seem to be part of some kind of academic cult.

    While such books can be informative and helpful, reading them can be akin to the slightly woozy feeling one has after having gone down a combined Google/Wikipedia/YouTube rabbit-hole, emerging with bewildered eyes and terabytes of regurgitated data. However, recent writing on the horror genre takes a different approach, eschewing the poles of either the popular or the academic for a perhaps yet-to-be-named third space. One book that takes up this challenge is Mark Fisher’s The Weird and the Eerie, published this year. (Fisher is likely known to readers through his blog K-punk, which had been running for almost two decades before his untimely death.) What Fisher’s study shares with other like-minded books is an interest in expanding our understanding of the horror genre beyond the genre itself, and he does this by focusing on one of the deepest threads in the horror genre: the limits of human beings living in a human-centric world.

    As a case study, consider the opening passage from H.P. Lovecraft’s well-known short story “The Call of Cthulhu”:

    The most merciful thing in the world, I think, is the inability of the human mind to correlate all its contents. We live on a placid island of ignorance in the midst of black seas of infinity, and it was not meant that we should voyage far. The sciences, each straining in its own direction, have hitherto harmed us little; but some day the piecing together of dissociated knowledge will open up such terrifying vistas of reality, and of our frightful position therein, that we shall either go mad from the revelation or flee from the deadly light into the peace and safety of a new dark age.

    With this – arguably the most foreboding opener ever written for a story – Lovecraft sets the stage for what is really an extended meditation on human finitude. Originally published in the February 1928 issue of the pulp magazine Weird Tales, “Cthulhu” ostensibly brings together the perspectives of deep time and deep space to reflect on the comparatively myopic and humble non-event that is human civilization – at least that’s how Lovecraft himself puts it. It is well known that Lovecraft took cues from the likes of Edgar Allan Poe, Algernon Blackwood, and Arthur Machen – influences that he himself notes. Equally well known is Lovecraft’s notorious xenophobia (often expressed in his correspondence as outright racism). Yet in spite of – or because of – this, Lovecraft remained unambiguous in his own approach to the horror genre. In his numerous essays, notes, and letters, he notes, with an unflinching misanthropy, how a horror story should evoke “an atmosphere of breathless and unexplainable dread of outer, unknown forces,” forces that point towards a “malign and particular suspension or defeat of those fixed laws of Nature which are our only safeguard against the assaults of chaos and the daemons of unplumbed space.” The “monsters” in such tales were far from the usual line-up of vampires, werewolves, zombies, and demons – all of which, for Lovecraft and his colleagues, end up serving as mere solipsistic reflections of human-centric hopes and fears. They are often described in abstract, elemental, almost primordial ways: “the colour out of space,” “the shadow out of time,” or simply “the lurking fear.”

    The story of “Cthulhu” itself  – which details the discovery of a cult devoted to an ancient, malefic, Elder Deity vaguely resembling a oozing winged cephalopod emerging from a hidden tomb of impossibly-shaped Cyclopean black geometry foretelling not only the end of the world but the deeper futility of the entirety of human civilization – the story itself has since obtained a cult status among horror authors, critics, and fans alike. In the early 20th century, like-minded tales of cosmic misanthropy were written by Lovecraft contemporaries Clark Ashton Smith, Robert E. Howard, and Robert Bloch, as well as by later authors of the weird tale such as Ramsey Campbell, Claitlín Kiernan, China Miéville, and Junji Ito. Like a slow-moving, tentacular meme, the Cthulhu “mythos” has reached far beyond the confines of literary horror. Film adaptations abound (the term “straight-to-video” no longer applies, but is still apt here). Video games, which nearly always end in despair and/or death. Role-playing games, complete with impossibly-shaped 10-sided black dice. A visit to any Comic Con will yield a dizzying array of comics, ‘zines, artwork, posters, bumper stickers, hoodies, Miskatonic University course catalogs, editions of the dreaded Necronomicon, and even Cthulhu plushies for the Lovecraft toddler. An industry is born. Today, distant cousins of Cthulhu can be seen in the Academy Award-nominated Arrival (2016), and the distinctly un-nominated burlesque that is Independence Day: Resurgence (2016). Cthulhu, it seems, has gone mainstream.

    Amid all the fondness for such abysmal and tentacular monstrosities, it is easy to overlook the themes that run through Lovecraft’s short tale, themes at once disturbing and compelling, and which mark the tradition often referred to as “supernatural horror” or “cosmic horror.” When Lovecraft characters happen upon strange creatures like Cthulhu (or worse, the Shoggoths), they don’t have the typical reactions. “Fear” is too simple a term to describe it; it encompasses everything without saying anything. But neither are they overcome by the more literary affects of “terror” or “horror,” like the characters of an old gothic novel. They have neither the time nor the patience for the critical distance afforded by a psychoanalytic “uncanny,” or the literary structures of the “fantastic.” Confronted with Cthulhu, Lovecraft’s characters simply freeze. They become numb. They go dark. Frozen thought. They can’t wrap their heads around that is right before them. What they “feel” is exactly this “inability of the human mind to correlate all its contents.” Forget the fear of death, I’ve just discovered a primordial, other-dimensional, slime-ridden necropolis of obsidian blasphemy that throws into question all human knowledge on this now-forsaken speck of cosmic dust we laughably call “our” planet.

    Yet, in all their pulpy, melodramatic, low-brow seriousness, the questions raised by Lovecraft and other writers in Weird Tales are also philosophical questions. They are questions that address the limits of human knowledge in a rapidly-changing world, a world that seems indifferent to the machinations of science or doctrinal exuberance of religion, impassive before the hubris of technological advance or the lures of political ideology – a cold “crawling chaos” lurking just beneath the fragile fabric of humanity. What the characters of such stories discover (aside from the usual train of madness, dread, and, well, death) is a kind of stumbling humbleness, the human brain discovering its own limit, enlightened only of its own hubris – the humility of thought.

    *

    This theme  – the limits of what can be known, the limits of what can be felt, the limits of what can be done – is central to Fisher’s The Weird and the Eerie. This is markedly different from other approaches to horror, which, however critical they may seem, often regard the horror genre as having an essentially therapeutic function, enabling us to purge, cope with, or work through our collective fears and anxieties. This therapeutic view of horror often becomes polarized between reactionary readings (a horror story that promotes the establishing or re-establishing of norms) or progressive readings (a horror story that promotes otherness, difference, and transgression of norms). And yet, in the final analysis, it is also hard to escape the sense that there is a certain kind of solipsism to the horror genre, that it is we human beings that remain at the center of it all, who have either constructed boundaries and bunkers and have once again staved off another threat to our collective identity, or who have devised clever ways of creating hybrids, fusions, and monstrous couplings with the other, thereby extending humanity’s long dreamed-of share of immortality.

    Whether reactionary or progressive, both responses to the horror genre involve a strategy in which the world in all its strangeness is transformed into a world made in our own image (anthropomorphism), or a world rendered useful for us as human beings (anthropocentrism). In spite of all the horrifying things that happen to the characters in horror stories, there is a sense in which the horror genre is ultimately a kind of humanism, a panegyric to the limitless potential of human knowledge, the immeasurable capacity for human feeling, the infinite promise of human sovereignty. This is, of course, not surprising, given the somber didactics of even the most extreme zombie apocalypses, vampiric mutations, or demonic plagues. Species self-interest is at stake. Humanity may be brought to the brink of extinction, only so that that same humanity may extend its mastery (self-mastery and mastery over its environment), and even obtain some form of ascendency over its own tenuous, existential status. Subtending the survivalist imperative of the horror genre and its pragmatic arsenal of mastering monsters of all kinds is another kind of mastery – a metaphysical mastery.

    But this is only one way of understanding the horror genre. The insight of books like Fisher’s is that the horror genre is also capable of chipping away at this species-specific sovereignty, taking aim at the twin pillars of anthropomorphism and anthropocentrism. Instead of being concerned with species self-interest and mastery, such horror stories tend more towards humility, hubris, and even, in its darkest moments, futility. It is a project that is doomed to failure, of course, and perhaps this why so many of the characters in the tales of Lovecraft, Algernon Blackwood, or Izumi Kyoka find themselves in worlds that are both untenable and unlivable. They end up with nothing but a bit of useless quasi-wisdom, scribbling away madly in a darkened forest room trying to make sense of it all not making any sense. Or they detach themselves from the humdrum human world of plans and projects, finding themselves inexorably pulled headlong into the ambivalent abyss of self-abnegation. Or worse – they simply continue to exist. What results is what we might call a “bleak humanism” – a horror story interested in humanity only to the extent that humanity is defined by its uncertainties, its finitude, its doubts – the humility of being human.

    Fisher’s terms are relatively clear. “What the weird and the eerie have in common is a preoccupation with the strange.” For Fisher, the strange is, quite simply, “a fascination for the outside […] that which lies beyond standard perception, cognition and experience.” But the weird and the eerie are quite different in how they apprehend the strange. As Fisher writes, “the weird is constituted by a presence – the presence of that which does not belong.” There is something exorbitant, out-of-place, and incongruous about the weird. It is the part that does not fit into the whole, or the part that disturbs the whole – threshold worlds populated by portals, gateways, time loops, and simulacra. Fundamental presumptions about self, other, knowledge, and reality will have to be rethought. “The eerie, by contrast, is constituted by a failure of absence or by a failure of presence. There is something where there should be nothing, or there is nothing where there should be something.” Here we encounter disembodied voices, lapses in memory, selves that are others, revelations of the alien within, and nefarious motives buried in the unconscious, inorganic world in which we are embedded.

    The weird and the eerie are not exclusive to the more esoteric regions of cosmic horror; they are also embedded in and bound up with quotidian notions of selfhood and the everyday relationship between self and world. The weird and eerie crop up in those furtive moments when we suspect we are not who we think we are, when we wonder if we do not act so much as we are acted upon. When everything we assumed to be a cause is really an effect. The weird and eerie are, ultimately, inseparable from the fabric of the social, cultural, and political landscape in which we are embedded. Fisher: “Capital is at every level an eerie entity: conjured out of nothing, capital nevertheless exerts more influence than any allegedly substantial entity.” There is a sense in which, for Fisher, the weird and the eerie constitute the poles of our ubiquitous “capitalist realism,” prompting us to re-examine not only presumptions concerning human agency, intentionality, and control, but also inviting a darker, more disturbing reflection on the strange agency of the inanimate and impersonal materiality of the world around us and within us.

    Fisher’s interest in Lovecraft stems from this shift in perspective from the human-centric to the nonhuman-oriented – not simply a psychology of “fear,” but the unnerving, impersonal calm of the weird and eerie. As scholars of the horror genre frequently note, Lovecraft’s tales are distinct from genre fantasy, in that they rarely posit an other world beyond, beneath, or parallel to this one. And yet, anomalous and strange events do take place within this world. Furthermore, they seem to take place according to some logic that remains utterly alien to the human world of moral codes, natural law, and cosmic order. If such anomalies could simply be dismissed as anomalies, as errors or aberrations in nature, then the natural order of the world would remain intact. But they cannot be so easily dismissed, and neither can they simply be incorporated into the existing order without undermining it entirely. Fisher nicely summarizes the dilemma: “a weird entity or object is so strange that it makes us feel that it should not exist, or at least that it should not exist here. Yet if the entity or object is here, then the categories which we have up until now used to the make sense of the world cannot be valid. The weird thing is not wrong, after all: it is our conceptions that must be inadequate.”

    *

    This dilemma (which literary critic Tzvetan Todorov called “the fantastic”) is presented in unique ways by authors of the weird tale and cosmic horror. Such authors refuse to identify the weird with the supernatural, and often refuse the distinction between the natural and supernatural entirely. They do so not via mythology or religion, but via science – or at least a peculiar take on science. In cosmic horror, the strange reality described by science is often far more unreal than any vampire, werewolf, or zombie. Fisher highlights this: “In many ways, a natural phenomenon such as a black hole is more weird than a vampire.” Why? Because the existence of the vampire, anomalous and transgressive as it may seem, actually reinforces the boundary between the natural order “in here” and a transcendent, supernatural order “out there.” “Compare this to a black hole,” Fisher continues, “the bizarre ways in which it bends space and time are completely outside our common experience, and yet a black hole belongs to the natural-material cosmos – a cosmos which must therefore be much stranger than our ordinary experience can comprehend.” Science, for all its explanatory power, inadvertently reveals the hubris of the explanatory impulse of all human knowledge, not just science.

    Authors such as Lovecraft were well aware of this shift in their approach to the horror genre. An oft-cited passage from one of Lovecraft’s letters reads: “…all my tales are based on the fundamental premise that common human laws and interests and emotions have no validity or significance in the vast cosmos-at-large.” To write the truly weird tale, Lovecraft notes, “one must forget that such things as organic life, good and evil, love and hate, and all such local attributes of a negligible and temporary race called mankind, have any existence at all.” So much for humanism, then. But Fisher is also right to note that Lovecraft’s tales are not simply horror tales. As Lovecraft himself repeatedly noted, the affects of fear, terror, and horror are merely consequences of human being confronting an impersonal and indifferent non-human world – what Lovecraft once called “indifferentism” (which, as he jibes, wonders “whether the cosmos gives a damn one way or the other”). There is an allure to the unhuman that is, at the same time, opaque and obscure. As Fisher writes, “it is not horror but fascination – albeit a fascination usually mixed with a certain trepidation – that is integral to Lovecraft’s rendition of the weird…the weird cannot only repel, it must also compel our attention.”

    This reaches a pitch in Fisher’s writing on author Nigel Kneale and his series of Quatermass films and TV shows. The Quatermass and the Pit series, for instance, opens with the shocking discovery of an alien spaceship buried within the bowels of a London tube station (which station I will not say). The strange, quasi-insect remains inside the ship point to another, very different form of life than that of terrestrial life. But the science tells them that the alien spaceship is actually a relic from the distant past. It seems that not only geology and cosmology, but human history will have to be rethought. Gradually, the scientists learn that the alien relics are millions of years old, and in fact a distant, early progenitor of human beings. We, in turns, out, are they – or vice-versa. The Quatermass series not only demonstrates the efficacy of scientific inquiry, it puts forth a further proposition: that science works too well. “Kneale shows that an enquiry into the nature of what the world is like is also inevitably an unraveling of what human beings had taken themselves to be…if human beings fully belong to the so-called natural world, then on what grounds can a special case be made for them?” Reality turns out to be weirder and more eerie than any fantastical world or alien civilization. This is what Fisher calls “Radical Enlightenment,” a kind of physics that goes all the way, a materialism to the nth degree, even at the cost of disassembling the self-aware and self-privileging human brain that conceives of it. Reversals and inversions abound. What if humanity itself is not the cause of world history but the effect of material and physical laws that we can only dimly intuit?

    This theme of Radical Enlightenment runs through Fisher’s book. While he does discuss works of fiction or film one would expect in relation to the horror genre (Lovecraft, Kubrick’s The Shining, David Lynch’s recent films), Fisher also offers ruminations on contemporary works (such as Jonathan Glazer’s 2013 film Under the Skin), as well as a number of evocative comparisons, such as a chapter on the weird effects of time loops in Rainer Werner Fassbinder’s film World on a Wire and Philip K. Dick’s novel Time Out Of Joint. There are also several surprises, including a meditation on the strange “vanishing landscapes” in M.R. James’s ghost stories and Brian Eno’s 1982 ambient album On Land. Also welcome is Fisher’s attentiveness to under-appreciated works in the horror genre, including the disquieting short fiction of Daphne du Maurier. In the span of a few carefully-written pages, Fisher follows the twists and turns of his twin concepts one chapter at a time, one example at a time, until it is revealed exactly how enmeshed the weird and the eerie are in culture generally.

    *

    The Weird and the Eerie is an evocative and carefully-written short study in cultural aesthetics. Far from the familiar line-up of vampires, zombies, and demons, Fisher’s eclectic examples speak directly to one of the central themes of the horror genre: the limits of human knowledge, the metamorphic shapes of fear, and the blurriness of boundaries of all types. His simple conceptual distinction quickly gives way to reversals, permutations, and complications, ultimately refusing any notion of a monstrous or alien unhumanness  “out there”; with Fisher, the unhuman is more likely to reside within the human itself (or as Lovecraft might write it, “the unhuman is discovered to reside within the human itself”).

    Many books on the horror genre are concerned with providing answers, using varieties of taxonomy and psychology to provide a therapeutic application to “our” lives, helping us to cathartically purge collective anxieties and fears. For Fisher, the emphasis is more on questions, questions that target the vanity and presumptuousness of human culture, questions regarding human consciousness elevating itself above all else, questions concerning the presumed sovereignty of the species at whatever cost – perhaps questions it’s better not to pose, at the risk of undermining the entire endeavor to begin with.

    I should let the reader decide which approach makes more sense, given the weird and/or eerie “Waldo-moment” in which we currently find ourselves. But the weird and the eerie are scalable, pervading broad cultural structures as well as the minutiae of personal ruminations. I’ve known Fisher as a colleague for some time. About a week after I had agreed to do this review, I heard via email of Fisher’s suicide. Someone I knew was previously there, over there, doing what they do, they way we so often presume a person’s presence in between moments of punctuated interaction. And then, suddenly, they’re not there. About a week after this, The Weird and the Eerie arrived in the mail. It was hard not to pick up the book and feel it had a kind of aura around it, as if it was some kind of final statement, a last communiqué. I had it on the table in a short stack with other books, and I kept half-expecting it to also vanish, as if its very presence there were incongruous. I would occasionally pick up the book and flip through it, as if secretly hoping to discover pages that weren’t there before. But my copy was the same as all the others. Besides, isn’t that essentially what a book is, a last word written by someone either long dead or who will die in the future? Maybe all books are eerie in this way.

    Eugene Thacker is the author of several books, including In The Dust Of This Planet (Zero Books, 2011) and Cosmic Pessimism (Univocal, 2015).

  • Ben Parker — What Is A Theory of the Novel Good For?

    Ben Parker — What Is A Theory of the Novel Good For?

    by Ben Parker

    Review of Guido Mazzoni, Theory of the Novel, translated from the Italian (2011) by Zakiya Hanafi, Harvard University Press, 2017.

    Because the novel is the most important product of modernity, any theory of the novel is also a theory of modernity. That modernity has been characterized in a variety of ways: as an unremitting catastrophe of Being—Georg Lukács’s The Theory of the Novel or René Girard’s Deceit, Desire, and the Novel; as the vulnerable legacy of humanist secularism—Erich Auerbach’s Mimesis; or epistemologically—Michel Foucault’s reading of Don Quixote as a crisis of signification. As Guido Mazzoni tells the story in his Theory of the Novel, modernity has been a long process of liberation from the implicit transcendence of collective cultural projects. We have now arrived at a moment where “the particular life represents the only horizon of sacredness that modern culture still recognizes.” Modernity is therefore the disruptive entropy of “unbelonging,” the triumph of “individualistic, anarchic, dispersive, centrifugal” forces over those of “collective transcendence.” By Mazzoni’s scorekeeping, the signal accomplishments of modernity are human rights, democracy, and relativism, but above all, “the concrete capacity to construct small spheres of autonomy.” The novel therefore marks “the entrance of democracy into literature,” because it is the vehicle par excellence of particularized private experience. Mazzoni prizes the novel for “its ability to make us see the world through the eyes and conscience of someone else, its ability to allow us to step into a possible life that is not ours.”

    Given this endpoint of absolute relativism—“Each person is an epicenter of absolute meaning”—Mazzoni has to construct his history of the novel retrospectively, as a gradual disburdening of the possibility of transcendence and collective horizons. He casts this ontological flattening in the light of an inner liberation of the novel form, although it could as easily be felt as a suffocating reduction. Mazzoni describes the first two centuries (1550-1750) of the novel’s history as an emancipation from the conceptual scaffolding of allegory and moral didacticism, on one hand, and from the strict delineations of classicist poetics (tragedy depicts a higher type of character, and comedy a lower) on the other. Because he was trained as a philologist, Mazzoni plunges the reader into a slough of terminological distinctions attending the birth of the novel: le roman, der Roman, il romanzo, romanice loqui, romanz, romance, novella, nouvelle, novela, novel. But his theory of genre rests upon a dubious metaphysics: rather than timeless Platonic forms, genres are “universals in re,” knots of emerging practices bound up with contemporary definitions and prescriptions. Instead of defining “the novel” retrospectively, which would mean fitting works like Tristam Shandy and The Golden Ass into the same Procrustean bed, Mazzoni sees the genre as the outcome of a complex fusion of heterogeneous conventions and literary corpuses. His approach is to “reconstruct the dialectic between the object and the words that enabled the object to be defined in the first place.” The drawback to this method is that the definition is never immanent to the novels themselves, but is derived from the belletristic scaffolding that is Mazzoni’s preferred archive. The scholarship on display—Mazzoni seems to have read every treatise and preface from the period—is unimpeachably exhaustive, even overwhelming. We learn that Don Quixote, for example, was not welcomed into the world as a novel but as a “comic romance.” But Mazzoni declines to pursue the question, what process of generic self-definition is Don Quixote itself engaged in? Nor does he see the retrospective genealogy of the novel as in large part an invention of the novel itself (as, for instance, the shelf of books in David Copperfield’s library). In any event, the upshot of this formative period is that the novel emerges as the “book of particular life,” a record of private persons, caught up in the “anarchy of the real,” rather than idealized or public figures made into abstract examples.

    Once the novel has broken free from allegory (whose political dimensions, overlooked by Mazzoni, have been detailed by Fredric Jameson), and we find ourselves in the nineteenth century, the next constraint to be discarded is melodrama. Melodrama gets painted as the bad outward form of psychology, which Mazzoni contrasts to the subtle analysis of interior life that culminates in James, Proust, and Woolf.  Thus melodrama turns out to be a convenient sorting mechanism for arriving at a set of all-too familiar preferences: Austen (but not Scott), Flaubert (but not Balzac), Eliot and Tolstoy (but not Dickens or Hugo). As with allegory, melodrama is classed as a transcendental and collective schema, averse to the finer gradations of “real life.” For melodrama, we are informed, belonged to a moment where “history had become a lived experience of the masses,” though “at a certain point this paradigm proved to be unrealistic.” It was no longer “plausible to think that people, subjects, or witnesses of an unprecedented transformation were involved in absolute conflicts.” What we have instead of large-scale history is the gradual extension of “our understanding of the interior life,” an ever-refined representational accuracy comparable to “the gains made in physics, astronomy, or anatomy.”

    By the time we reach the contemporary novel, the sphere of freedom that Mazzoni wants to find in the novel has been narrowed down to the horizon of sheer everydayness. We have exchanged the wild explorations of Robinson Crusoe, Gulliver, Edward Waverley, Natty Bumppo, and Huckleberry Finn for the boredom of Emma Bovary. All we are left with is the bad infinity of “real life” in its banal givenness. Freedom is surreptitiously redefined, from the kind of “unbelonging” of the earlier mode of “lighting out for the territory,” to the unbelonging of grousing individual discontent. No surprise that the contemporary authors Mazzoni endorses are Philip Roth, J.M. Coetzee (singling out Boyhood and Youth), Michel Houellebecq (The Elementary Particles), and Jonathan Littell. He doesn’t provide a reading of any of these novels (although he does cite a negative review of Littell’s The Kindly Ones). Knausgaard’s novel is something like an empirical confirmation of Mazzoni’s thesis about the tendency of the novel towards absolutely private particularity, absent any transcendent justification. Mazzoni’s concluding observation—“Inside our small local worlds, everything at stake has an unquestionable value” —could just as easily have been written by Knausgaard as a summary of the exhausting strife of representability at the heart of his book.

    In outline, then, Mazzoni’s account recapitulates the problematic of Lukács’s Theory of the Novel—“the refusal of the immanence of being to enter into empirical  life,” the pulverization of all transcendent projects—in order to render it unproblematic. What Lukács saw as “the dissonance special to the novel” was its capturing of the devastating ironies and grotesque realizations that the transcendent ideal is exposed to. For Mazzoni, however, such dissonance is simply “implausible,” a failure of perspective insufficiently immersed in the proliferating contingencies of “real life.” So, what for Lukács was the constitutive problematic of the novel—the hard-fought contest between the ideal and an inert (but ultimately victorious) reality—here turns out to be a detachable “extra” or a historical vestige. Mazzoni sees the struggle with the ideal as something that was gradually exorcised or shed during the novel’s development, as opposed to something essential to defining the genre. His argument then turns out to be another entry in the “end of grand narratives” narrative, or an instance of what Alain Badiou calls “democratic materialism”: we no longer believe in any Truths striving to be realized in the world, only in local particulars. With oracular resignation, Mazzoni announces that, starting with some generalized metaphysical eclipse in the nineteenth century, “Universal forces were no longer revealed in the experience of private persons.” One imagines him lecturing the great characters of fiction like a stern guidance counselor, for their stubborn lack of realism, in those moments of Lukácsian “dissonance” where they confront a churning abyss of unbearable meaning underlying an ongoing and inessential life: Don Quixote for attempting to revive chivalry by mounting his gaunt nag and donning a pasteboard visor; or Catherine Earnshaw for proclaiming, “I am Heathcliff!”; or Captain Ahab for hurling himself against the whale as striking at some “inscrutable malice” behind a mask; or Marlow for detecting, in the depths of the Congo, “the stillness of an implacable force brooding over an inscrutable intention.”

    To be sure, Mazzoni’s claim that the novel has freed itself from the transcendental has the force of self-evidence, if one surveys contemporary fiction. Mazzoni’s reading of novels in English cuts off at 2002, but (in addition to Knausgaard) Chris Kraus, Sheila Heti, Ben Lerner, and Rachel Cusk would all be pertinent here, as instances of flattened, quotidian perception, where the “microcosm” of private existence—voided of melodrama or narrative artifice—is elevated to “absolute importance.” Going further back, one could add other instances. John Updike, Frederick Exley, and Renata Adler come immediately to mind. Mazzoni doesn’t mention Norman Mailer, who is on quite another track, but whose “nonfiction novel” would be additional confirmation of the novel’s tendency to represent a reality divested of transcendent impulses. (At this point, however, one wonders whether it were not fictionality itself that represents the final burden of transcendence, whether Mazzoni’s sense of “the novel” is not just headed towards the documentary status of journalism, memoir, travel writing, etc.)

    On the other hand, some of the most acclaimed novels of recent years have resuscitated either melodrama (Hana Yanagihara’s A Little Life), or transcendental (religious) preoccupations (Marilynne Robinson’s work), or allegory (Yann Martel’s Life of Pi). To remark these works are also somewhat middlebrow and embarrassing, would introduce a dimension of aesthetic evaluation that Mazzoni never broaches. It’s worth noting, too, that Mazzoni’s own examples are not unproblematic. Although Houellebecq’s The Elementary Particles does duty for Mazzoni, his more recent The Possibility of an Island and Submission don’t fit the pulverization-of-collective-transcendence thesis at all. Houellebecq emerges, instead, as an (unevenly satirical) utopian thinker, closer to Jonathan Swift in the Houyhnhnms section of Gulliver’s Travels than to Roth’s Zuckerman novels. Mazzoni also cites the autobiographical novels of J.M. Coetzee, but his latest novels, The Childhood of Jesus and The Schooldays of Jesus, whatever else they may be, are obvious violations of Mazzoni’s rule against allegory.

    The unbearable scene he cites from Buddenbrooks, when little Hanno draws two lines under the last entry in the family tree, muttering, “I thought… I thought… there wouldn’t be anything more,” is indeed a powerful image of finitude. But Mann then went on to write the highly allegorical The Magic Mountain and Doctor Faustus. Dostoevsky is invoked in a number of contradictory ways—he is, on one hand, one of the first authors who is “still contemporary,” because of his techniques of characterization, but on the other hand, he presents a regrettable and lingering case of melodrama. What is never mentioned is that Dostoevsky’s oeuvre, from start to finish, is rent through with transcendental preoccupations. To take only the case of The Brothers Karamazov, what does one make of the beautiful moment in the final chapter, where the father of the slain child Ilyusha sees a flower fall on the snow, and rushes “to pick it up as though everything in the world depended on the loss of that flower”? This sense of absolute responsibility, of “everything in the world” depending on one’s posture towards salvation and loss, is the hard core of Dostoevsky’s meaning. If Mazzoni wants to insist that “we cannot go beyond” our immersion in factical being, that it is “the sole layer of existence that… distinguishes us from nothing,” then he will have to lose The Brothers Karamazov as a forward-looking work.

    I wrote above that the novel is the most important product of modernity. I forgot to add that modernity is in large part the product of the novel. The novel is one of the “workshops where ideals are manufactured,” to take an image from Nietzsche’s Genealogy of Morals. For instance, the continuous and rigorous thinking of responsibility throughout the novels of the Victorian period (paradigmatically, Great Expectations, Tess of the D’Urbervilles, and Lord Jim) constitutes as central a development of our ethical life as the subsequent Freudian theorization of same. The self-representation of the nineteenth-century social imaginary is largely created through the ways novels develop of “giving an account of oneself,” in Judith Butler’s phrase. The ultimate trouble with Theory of the Novel is that Mazzoni oscillates between seeing the novel as a co-creator of modernity, whereby “an essential aspect of the Western form of life takes shape and becomes an object of knowledge only through mimesis and fiction,” and seeing the novel (or cultural production as a whole) as validating (or falling into line with) larger systemic results, e.g. “the disintegrative force implicit in modern individualism,” or “the relativistic deflation of collective values.” We don’t know, finally, whether the Western “crisis of transcendence”—what for Lukács was an ongoing schism constitutive of the novel form—is simply a fait accompli restricting literary possibility, or whether one might hold the history of the novel itself accountable for this disintegration. Nor does Mazzoni see the novel as a possible reflection upon these outcomes, a perspective-taking that would refuse the enforcement of deflationary relativism.

    But might not the greatest novels be precisely such refusals? To return again to The Brothers Karamazov, we find there (in the remembrances of Father Zosima) a forestalling of Mazzoni’s conclusions, in almost identical language: “For all men in our age are separated into units, each seeks seclusion in his own hole, each withdraws from the others, hides himself, and hides what he has, and ends by pushing people away from himself… He is accustomed to relying only on himself, he has separated his unit from the whole, he has accustomed his soul to not believing in people’s help, in people or in mankind.” For Dostoevsky, at least, the novel is not a story of emancipation from transcendence. If the novel has nevertheless brought about this anomie and purgation of values, the novel goes on only in a perpetual fight against what it hath wrought.

    Ben Parker is assistant professor of English at Brown University. His current research is on recognition scenes in the nineteenth-century novel. He tweets @exyoungperson.

  • Sarah Brouillette — Couple Up: Review of “Family Values: Between Neoliberalism and the New Social Conservatism”

    Sarah Brouillette — Couple Up: Review of “Family Values: Between Neoliberalism and the New Social Conservatism”

    by Sarah Brouillette

    Review of Melinda Cooper, Family Values: Between Neoliberalism and the New Social Conservatism (New York: Zone Books, 2017)

    The basics of neoliberalism are by now well known. Pressured to be wary of public deficit spending, and trying to find ways to rejuvenate depressed economies, neoliberal governments cut spending on welfare and other social services, and turn the programs that do remain into job training “workfare.” Policies at the same time shift to give priority to the needs of businesses wanting to keep wages low, to offshore production, and to make few or no commitments to workers. The power of unions is undercut as a result, so it is decreasingly possible to look to that form of collectivity as a shelter.[1] Politicians, advisors, sympathetic management consultants and business professors meanwhile emphasize private initiative and personal merit as the keys to success. As a result, work has been trending toward the less regular, less routine, less secure, less protected by union membership, with wages stagnant and less likely to be supplemented by things like affordable public education, low rents, tax credits, and childcare benefit payments.

    The working individual suited to this environment will naturally possess certain traits, as people are encouraged to look to themselves for more and more of what they need. Everything becomes a matter of personal responsibility: invest smartly for the future, take out a loan to pay for college, be your own brand, find your joy, “live your life.” If there is a culture of neoliberalism, it is all about interiority and the individual psychic life: therapeutic culture, because there is little state funding for mental health treatment. Find out who you really are, do what you love, look within, take your natural resilience as the base of every struggle and its overcoming; experience setbacks, Pop Idol style, as welcome occasions to overcome every hurdle. Self-improve. Self-actualize.

    The causal relations are sometimes murky and eminently debatable. Don’t governments in fact fund wellness initiatives, especially targeting underprivileged communities? And what about all the counternarratives emphasizing the necessity of communities coming together – the British Tories’ “Big Society,” for instance? But the general account of neoliberalism is quite uniform. It pinpoints the force of biographization, responsibilization, individualization, self-management, a DIY ethos, and customization of personal preference as the lifeblood of the neoliberal order.[2]

    Against all this, Melinda Cooper’s Family Values: Between Neoliberalism and the New Social Conservatism argues that the key social unit of neoliberalism is not the individual but the family, and not just any family but the family in perpetual crisis. She presents the postwar Fordist family wage – basically, a state-backed wage high enough to support a family with only one parent working – as a “mechanism for the normalization of gender and sexual relationships” (8), and for this reason sees no reason to lament its demise. As an “instrument of redistribution,” she writes, it “policed the boundaries between women and men’s work and white and black men’s labor” and was “inseparable from the imperative of sexual normativity” (23). “Few African American men enjoyed the family wage privileges of the unionized industrial labor force,” and their disproportionately high unemployment is evidence of the “multiple exclusions serving to define the boundaries of state-subsidized reproduction” (35-6).

    Just as the “Fordist politics of class … established white, married masculinity as point of access to full social protection” (23), the fundamental concern for neoliberals like Gary Becker was how to respond to the breakdown of this masculinity and the family built around it. “Neoliberals are particularly concerned about the enormous social costs that derive from the breakdown of the stable Fordist family,” Cooper argues. They aim “to reestablish the private family as the primary source of economic security and the comprehensive alternative to the welfare state” (9). Basically, they want the traditional family intact as a compensation for precarity.

    The data show that in the neoliberal era private family wealth is increasingly decisive in “shaping and restricting social mobility” (125), and this is a result of concerted policymaking. In the 1960s, inflation eroded the wealth at the top tiers, as it translated into the deflation of financial assets. Inflation was at the time understood in precisely this way, as a redistributive tax, “intensifying progressive tendencies” of the period: “Free-market economists insinuated that inflation was a form of state-sanctioned fraud – a covert tax designed to extort wealth from investors and transfer is to the lower classes” (127). The neoliberal “paradigm shift in American fiscal and monetary policy” sets about ending this redistributive movement. If the Employment Act of 1946 wanted to “promote maximum employment, production and purchasing power,” where wage and price inflation were understood as signs of growth and as “benign trade-offs to full employment,” the neoliberals overturned all this.

    Figures such as Milton Friedman and Paul Volcker “turn[ed] inflation-targeting into the prime objective of monetary policy,” thus restricting the money supply and pushing up interest rates. Whereas bondholders in the 1970s saw assets depreciate and the Federal Reserve “deferred to the interests of unionized labor and welfare constituencies,” in the new era the Fed would strive “to repress wages and consumer prices in the service of asset price appreciation.” These policies led to a sure turnaround in the distribution of national income; the “share of national income flowing to financial investors went from negative or stagnant in the 1970s to ‘substantially positive’ in the 1980s”; while “labor’s share of national income declined proportionately” (134). By 1983, Cooper writes, “wealth concentration had reverted to its 1962 level and by the end of the decade had plummeted to levels comparable to 1929” (135).

    There has thus been, at the top tier, a massive “resurgence of large family fortunes” (137). Nearly everywhere else, though, with stagnant wages, unemployment, and the transfer of the costs of things like higher education and health care back to families, lack of access to familial wealth can condemn one to a lifetime of debt. Hence Cooper’s argument about the importance of the family: intergenerational familial support in the form of housing, or money, or willingness to be signatories to loans, is a neoliberal necessity for many, and the pressure to combine dependence on parents with married coupledom just compounds the effect.[3] According to statistics gathered by the Pew Research Center, 1960 was the year in which people under 25 were most likely to live independently. In more recent decades, however, young people have been exhorted to invest in the future, save for retirement, and acquire assets (houses and university degrees). At the same time, and often in relation to this, they have been forced into debt and into insecure employment. No wonder they are more inclined to live with parents or partners. Of course, there is such a thing as a non-normative family, and perhaps living independently from relatives is not something we should unduly idealize. Cooper’s interest, though, is in what sort of family arrangements government programs prefer, and how preferences shift given combined pressure from neoliberal economic policy and the new social conservativism. We will return to her idealization of independence, however.

    The more common argument, of course, is that neoliberalism is destructive to family life, as it encourages workers to be “low drag,” moveable, flexible, always working, losing any sense of a private life outside of work, and also alone in leisure in front of a personally selected entertainment service displayed on a privately watched device. Yet, as Annie McClanahan has recently argued, not many people are really these footloose mobile workers.[4] For most employers, it is probably more important that those they hire be replaceable than that they be mobile. Only workers in relatively elite sectors (high tech, higher education, entertainment) are in a better position if they can move from thing to thing without worrying about family obligations.[5]

    This is not to deny that there is now also a more general animus against the restrictions and burdens of family life – the boredom of marriage, and drudgery of raising children (all captured so well by a show like Mad Men, for instance, which crystalizes the individualizing ethos so perfectly). However, there is just as much pressure to maintain the bonds of coupledom, and this tension between rejection and embrace may in fact be the point worth emphasizing. It seems that people are increasingly wondering if marriage is “worth it,” while decreasingly being able to exit it, and this is a cause of general anxiety, finding outlet in things like the dating site for adulterers, Ashley Madison, which was notorious for a minute in 2016 after its user data was stolen. When it turned out that most of the male customers were at least some of the time corresponding with bots rather than real women, I couldn’t help thinking that in a way it didn’t matter: the point is that users find an outlet for their sense of being stuck in a social relation (marriage!) on which they are dependent. Indeed, the bot’s lack of reality, lack of availability, is what makes the “affair” appealingly nonthreatening to the user’s IRL relationships. Moralistic attacks on these men – the fact that some of those caught are family-values conservatives is, to be sure, a rich irony – miss the point: they are not having affairs; they are staying in unhappy marriages that they depend on in various ways.

    They depend on marriage because it is still the normative standard for people (if you aren’t married there is something wrong with you; if you don’t have kids you are deviant in some way). They depend on it in that they can’t afford a house without two salaries, because for tax purposes it is better to be a legally recognized couple, because the lifestyle they aspire to requires it, because caring for children alone is very hard, because shifting work hours and temporary contracts make the second salary a necessity, even if it too is precarious. They depend on it because they are too tired and generally physically weary to try to have any other sort of relationship. Being non-normative can feel like SO. MUCH. WORK. A film like 2009’s Up in the Air makes the point very well: the protagonist is the epitome of the roving high-powered executive entrepreneur (indeed, his job is to fire people), but his story is not a celebration of the escape from normativity. It is rather a lament about the psychic misery of solitude. The message is clear: couple up!  

    How did the family start to lose its normative power? For Cooper, conservatives skewering feminism, and more leftist thinkers trying to understand the foundations of neoliberalism, are in agreement about the force of 1960s and 1970s countercultural and antinormative critiques of the family. In Wolfgang Streeck’s analysis, the revolution in family law and intimate relationships – for example, the availability of no-fault divorce – destroyed the Fordist family wage because women were not stuck in the kitchen dependent on men any more. The family became a more flexible form because, in Cooper’s paraphrase of Streeck, feminists sought “an independent wage on a par with men,” eventually “transforming marriage from a long-term, noncontractual obligation into a contract that could be dissolved at will” (11). Cooper reads Eve Chiapello and Luc Boltanski’s argument as similar, in that they show how “the artistic left prepared the groundwork for the neoliberal assault on economic and social security by destroying its intimate foundations in the postwar family” (12). She quotes Nancy Fraser, also, who has written that “critique of the family wage … now supplies a good part of the romance that invests flexible capitalism with a higher meaning and moral point” (12). In each case, the idea is that feminism is somehow to blame for neoliberalization, because in seeking to free women from certain kinds of normative obligation and dependency they have demonized dependency in general, fetishizing independence from supports of any kind. Against these analyses, Cooper asks: what breakdown of the family, anyway? The apparent post-normativity of contemporary life is entirely compatible with the establishment of new norms. We continue to be form-determined after we no longer see social forms’ normative force. Put simply: the traditional family, which for Cooper is a family coerced into existence by exigency and normativity, is not broken enough.

    The economy in depression no longer affords the state-supported Fordist wage, but the family is re-inscribed and reformulated even as it is queried and undermined by antinormative movements. If the foundations of neoliberal policy are thoroughly economic, neoconservativism enters Cooper’s account as a largely compatible reaction formation. The neoconservative agenda, formed deliberately against the liberation movements of the 1960s and their challenge to the normativity of the traditional family, served neoliberalization far more than the countercultural left’s challenges to social convention. Cooper argues that, whereas nostalgia for the Fordist wage became a “hallmark of the left,” neoconservatives, allied with thrifty neoliberals, preferred “the strategic reinvention of a much older, poor-law tradition of private family responsibility.” In a policy formation that reflected both neoliberal and neoconservative thought, social welfare was not to disappear, but instead to be made into “an immense federal apparatus for policing the private family responsibilities of the poor” (21).

    As a public assistance program targeted at the noncontributing poor – workers paying into funds that would support them in the event of unemployment were always more palatable (34) – the fate of AFDC (Aid for Families with Dependent Children) is one of Cooper’s main cases. It allows her to show how social welfare extended to the poor – especially to single women, especially mothers, especially black mothers – became “associated with a general crisis of the American family” (29). As the composition of the program changed, with the number of African American women signing up outpacing that of white woman, and divorced or never-married women joined the rolls, fears were heightened. Because “racial and sexual normativities were truly foundational to the social order of American Fordism, determining just who would be included and who would be excluded from the redistributive benefits of the social wage” (36), the inclusivity evident in the 1960s in the AFDC’s provision for non-married mothers proved to be short-lived. Arguments for reinstating the stability offered by the traditional family had significant influence at this juncture.

    Nor were these arguments solely made by conservatives. In the 1960s there was in fact significant leftist promotion of the African American male-breadwinner family and a related impetus against “non-normative lifestyles of unattached African American women” (37); hence the tendency to identify the AFDC as a cause of family breakdown while promoting the “male breadwinner’s wage” (41). An article by Richard A. Cloward and Frances Fox Piven, published in The Nation in 1966 and presented as “a strategy to end poverty,” laments that the state was “substituting check-writing machines for male wage earners,” thereby “robb[ing] men of manhood, women of husbands, and children of father.” The authors continue: “To create a stable monogamous family, we need to provide men (especially Negro men) with the opportunity to be men, and that involves enabling them to perform occupationally” (qtd. 42).

    What they saw were “perverse disincentives to family formation built into the AFDC program” (43), whereas women left more to their own devices would naturally be more likely to find men to support them. With the 1970s economic downturn, and anxieties directed at inflation in particular, the program became a touchstone for debates for neoconservatives formulating their “new political philosophy of non-redistributive family values” (47). While neoliberals “called for an ongoing reduction in budget allocations dedicated to welfare—intent on undercutting any possibility that the social wage might compete with the free-market wage,” neoconservatives advocated an expanded role for state in regulating sexuality. On both fronts, the point was the urgent necessity of “reinstating the family as the foundation of social and economic order” (49).

    Cooper discusses Milton Friedman’s concern that the “natural obligations” that “once compelled children to look after their parents in old age” have given way to “an impersonal system of social insurance whose long-term effect is to usurp the place of the family” (58). Friedman wrote that whereas once “Children helped their parents out of love or duty,” they now “contribute to the support of someone else’s parents out of compulsion and fear” (qtd. 58). State-based redistribution was a poor substitute for proper familial support and wealth transmission. For Gary Becker, also, the postwar welfare state destroys the “natural altruism of the family” (60). Becker’s theory of human capital is perhaps the premier theorization of individual self-management and self-appreciation. Michel Foucault treated Becker’s work as exemplary of the way that neoliberal analyses entail “replacement every time of homo economicus as partner of exchange with a homo economicus as entrepreneur of himself, being for himself his own capital … a capital that we will call human capital inasmuch as the ability-machine of which it is the income cannot be separated from the human individual who is its bearer.”[6] Becker also featured recently in a Merriam-Webster tweet of the term “human capital” – “turning people into statistics since 1799,” the tweet quipped – which linked to the full dictionary entry, where Becker’s work is cited as “taking a holistic view of a person’s life and experiences as they can be applied within the workforce.” Becker took personal investment in one’s own human capital appreciation as preferable to state investment (the benefits of high human capital only accruing to oneself, after all), and thus supported rising tuition costs and the student loan industry as a major part of the growing importance of private credit. Yet Cooper shows that his arguments also preferred a supportive wealth-generating family: the older generations would back student loans where necessary, as they naturally want children and grandchildren to bear human capital that self-appreciates at a greater pace and with results that are more lucrative. Becker celebrated Ronald Reagan for restoring kinship bonds.

    Reagan drastically cut the AFDC, before Bill Clinton eliminated it. It was replaced with the TANF program (Temporary Assistance for Needy Families), whose availability was contingent on states’ willingness to track down and enforce paternity obligations. TANF’s defenders claimed it is better for a woman and her children to be reliant on alimony and child support than to turn to the government for assistance (67). Here we get to the heart of Cooper’s refutation of the idea that neoliberalism privileges the footloose free agent. In fact, in her account, neoliberalism is more likely to pressure people to sustain unhealthy and unsustainable family and intimate relationships, including tying children to fathers who do not know them or want them. Clinton’s extensive welfare reform reflected and codified what she calls “a new bipartisan consensus on the social value of monogamous, legally validated relationships.” His government reformed welfare spending while devising “initiatives to promote the moral obligations of family, including a special budget allocation to finance marriage promotion programs and … bonus funds to states that could demonstrate that they had successfully reduced illegitimate births without increasing the abortion rate” (68). Barack Obama’s “healthy marriage and responsible fatherhood” initiatives continued in this direction.

    Cooper suggestively connects these initiatives to the “first experiment in federal relief ever implemented by Congress”: the 1865 creation of the Freedmen’s Bureau following the Emancipation Proclamation of 1863. Before 1863, slaves were precluded from legally sanctioned marriage. The Freedman’s Bureau instructed that freedom to participate in the labor market came with “the right to marry and the responsibility to support wife and child” (79). Its support for freed slaves entailed a vigorous campaign to promote marriage, with Bureau agents authorized to perform marriages and a “sustained pedagogy of domestic life, schooling men in the notion that they were to become the breadwinners of the family and women in a new kind of economic dependence” (80). There were penalties for people cohabiting without marriage; and Bureau-assigned wage scales that penalized women, precisely because of the “social costs of dependency” that fell upon the state if forced to support unmarried women and their children (81).

    Like the more recent insistence that women secure alimony and child support before turning to welfare, these policies empowered men to assert rights over women and children. Indeed, the assumptions upon which they were based were not fundamentally threatened until the 1960s liberalization of family law, which made divorce easier and eased the stigmatizing of non-marital unions and cohabitation. “For an all too brief moment,” she argues, “revised AFDC rules allow divorced or never-married women and their children to live independently of a man while receiving a state-guaranteed income free of moral conditions” (97). That moment is over, however. “The modern child support system serves to demonstrate that the state is willing to enforce—indeed create—legal relationships of familial obligation and dependence where none have been established by mutual consent,” Cooper writes (105).

    We should pause here now on the figure of the never-married woman living independently thanks to welfare. Cooper argues that, in a context of relatively healthy public welfare spending, and of the pressures put on states by countercultural and antinormative activisms, there was a time when social welfare was “making women independent of individual men and freeing them from the obligations of the private family” (97). Hence, the fuel for the neoconservative backlash that soon followed – a backlash that gained traction because of the failing economy to which neoliberals were also turning their attention. A perfect storm. Yet Cooper’s celebration of the period in which social welfare possibly freed women from the constraints of marriage has her falling back into the trap she dismantles elsewhere: nostalgia for state provision.

    The image of the single woman with children, living with a state-based income “free of moral conditions,” reads as an idealization. Certainly, supporting children as a single parent on welfare has never been a cakewalk; and, are we meant to conclude that “freeing” men from the burdens of paternity is an unalloyed boon to women? She needs this figure, though. Cooper’s idealization of the state-supported single mother alerts us to the fact that her ultimate objection is not to social welfare but rather to the restriction of its benefits to the Fordist white male breadwinner, and to the way welfare programs get tied to normative policies and programs emphasizing the preferability of turning to family, especially marriage, to marshal the necessary resources to get by.[7] She avoids the stronger critique of social welfare, which might emphasize the global accumulative regime and resource extraction on which US prosperity was built, how nation-based welfare disperses the benefits of prosperity to some and not others, and the welfare state’s various regulatory and pacifying functions.[8]

    Does neoliberalism feel different to some people simply because it follows on the moment of postwar prosperity and the relatively expansive Keynesian social welfare that flowed from it, in which there was palpable faith in the civic virtue attending government spending on social programs? Neoliberal policies have threatened protections and comforts that these programs offered to some people – people like American and British university professors, who produce the analyses of the unique wrongs of the neoliberal order. Is all the worry about neoliberalism just a symptom of the decline of the hegemony of liberal democracy?

    The economy that supported the pre-neoliberal era of relatively high wages, and relatively generous public deficit spending on welfare and education, was also hugely resource extractive and suburbanizing. The capacity to redistribute wealth more evenly in the US was, in addition, contingent upon broader economic transformation that required dispossessions, expulsions, enclosures, primitive accumulations, US hegemony propped up by global wars, and the origins of the whole phenomenon of US industrial triumph after WWII in wartime accumulation and relative devastation across Europe.[9] Wherever one looks, the accumulation of wealth requires these devastations, making even the lushest times at the ADFC, and the possibility for a temporary flourishing of alternative kinds of family structures, into a troubled gain. For these reasons, it may be that work that avoids the terminology of neoliberalism, or uses it warily – work by Endnotes, by Silvia Federici, or by Robert Brenner, for instance – provides better purchase on contemporary conditions. Because when they fail to name the fundamental, global, totalizing causes of policy shifts, accounts of neoliberalism miss the ruthlessly intensifying dynamics of capital accumulation that are simply propelled onward with extended credit.[10]

    Finally, if Keynesian social welfare is a wage supplement designed to encourage consumer spending, in which sense is it wise to pit it against the dominance of commerce and private interests? If extensive public deficit spending on social programs and neoliberal monetarism are just different ways of managing the economy, and if one takes the capitalist economy as fundamentally anathema to universal human flourishing, to what extent should we worry about the difference that neoliberalism makes? Family Values doesn’t quite answer these questions. However, it does do the crucially important work of historicizing the rise of private credit in relation to family-values conservativism, and dismantling the left-liberal tendency to lament neoliberalization because it clawed back the gains of the immediate postwar period. Without suggesting that no gains were made, Cooper shows how they were thoroughly mitigated by normative racial, sexual and gender ascription – ascription that determined how to divvy up Fordism’s generous provisions, and that continues to push people, especially the already suffering, into unwanted contracts in life and work.

    Notes

    [1] For a recent account along these lines see Wendy Brown, Undoing the Demos: Neoliberalism’s Stealth Revolution (New York: Zone Books, 2015).

    [2] See for instance Ronen Shamir, “The age of responsibilization: on market-embedded morality,” Economy and Society (37.1: 2008): 1-19.

    [3] I discuss Cooper’s blistering account of the student loan industry elsewhere.

    [4] Annie McClanahan, “Becoming Non-Economic: Human Capital Theory and Wendy Brown’s Undoing the Demos,” Theory & Event 20.2 (2017): 510-519.

    [5] Even scholars suggesting that, in being less interested in keeping people in regular work, crisis-era capitalism allows for “queer liberation” from cis-hetero norms, insist in the next breath that some elements of queer life are tolerable and easy assimilated – think pink washing and gay marriage – and some are not.

    [6] Michel Foucault, The Birth of Biopolitics:  Lectures at the Collège de France 1978-1979, trans. Graham Burchell (Palgrave, 2008): 226.

    [7] In an earlier work, where the figure of the state-supported single mother is absent, her take is more ambivalent. She argues that the welfare state “undertakes to protect life by redistributing the fruits of national wealth to all its citizens, even those who cannot work, but in exchange it imposes a reciprocal obligation: its contractors must in turn give their lives to the nation” (Melinda Cooper, Life as Surplus: Biotechnology and Capitalism in the Neoliberal Era [University of Washington Press, 2008]: 8).

    [8] Gavin Walker has recently argued that “the function of ‘welfare’ within capitalism has never been something separate from its workings; rather, it is something co-emergent and central to the operation of the capital-relation itself”: “Rather than being a political development in which capital’s violence is ameliorated through social spending, we should rather understand the welfare state as the primary mechanism through which the process of primitive accumulation can be continuously sustained in the advanced capitalist countries” (“The ‘Ideal Total Capitalist”: On the State-Form in the Critique of Political Economy,” Crisis & Critique 3.3 [2016]: 434-455).

    [9] For an account along these lines see “Misery and Debt,” Endnotes 2 (April 2010): 20-51.

    [10] I owe this point to discussion with Tim Kreiner.

  • Naomi Waltham-Smith — Review of “Sonic Intimacy: Voice, Species, Technics (Or, How to Listen to the World)”

    Naomi Waltham-Smith — Review of “Sonic Intimacy: Voice, Species, Technics (Or, How to Listen to the World)”

    by Naomi Waltham-Smith

    Review of Dominic Pettman, Sonic Intimacy: Voice, Species, Technics (Or, How to Listen to the World) 

    What if the world had a voice? What would a world suffering under the burden of human dominance over the environment—what would that geological epoch known as the Anthropocene—say to us? Dominic Pettman asks us to imagine such a world in which not just human beings or animals but all living and inanimate objects, and even virtual technologies have voices. Sonic Intimacy invites us to tune into the seductive voice of an OS in Spike Jonze’s 2013 film Her, the swansong of the Sirens, the meowing of a cat, the melancholy songs of a lonely whale, the wind in the trees, even “the imploring squeal of a garden gate, crying out for oil” (49). This is a world in which listening, too, is not confined to human ears. In Pettman’s book, listening is even extended beyond the animal world in a range of examples both banal and symbolic: if mothers listen to their daughters’ voices on the phone and dogs to His Master’s Voice on the gramophone, lamps also prick up their ears at the clap of a hand and microphones listen for algorithmically determined shapes in order to identity specific words or even voices.

    Pettman’s call to hear those other voices and thus become those other kinds of listeners stems to no small degree from our deafness to what is arguably the greatest threat the world faces today and to the human and ecological crises that climate change is already precipitating. “Alarmed scientists try to tell us on a daily basis,” Pettman points out, “that we are not listening to the earth, which is—elliptically perhaps, and in its own cryptic way—trying to tell us that it is in trouble” (6–7). He argues that in the ongoing calamity that is the Anthropocene, it is vital that we challenge anthropocentric constructions of the voice and of the ear. If there is one main target in Sonic Intimacy, it is human exceptionalism. This critical outlook has shaped Pettman’s work in post-humanism more generally. For instance, the recent Creaturely Love observes how the images of human desire we construct tend to disavow our own animal natures.[1] Pettman’s earlier Human Error (published in 2011) explored mistaken efforts to define humanity in its opposition to machines and instead posits a cybernetic triangle of human, animal, and machine so as to decenter the human.[2] Humanity’s species-being, as he argued in that book, had become “specious-being,” not simply a mistaken identity, but the mistake of identity.

    Each of Sonic Intimacy’s four chapters explores a voice that is, if not post-human, in some way more or less than human—a negation of the human. The first, devoted to the voices that speak to us from machines, centers on a discussion of Jonze’s Her, in which a heart-broken man falls in love with his operating system “Samantha.” The film illustrates that bodies do not simply produce voices; conversely, voices can also produce bodies. As an awkward scene in which Samantha ventriloquizes the body of a mute stranger shows, acousmatic voices can be more involving and erotic than actual bodies. In this way Pettman establishes the idea of a sonic intimacy that is intimate precisely in having shed its physical presence. This observation leads Pettman to seek to explain the absence of “aural porn” on the internet (yes, dear reader, such are the surprising twists and turns of this riveting book!). If the voice, untethered from the overdetermined female body, were allowed to circulate unchecked, it would threaten the entire patriarchal system—a system that depends precisely on the exclusion and capture of an inarticulate cry consistently coded as female or animal. Hence—paving the way for the next chapter on the gendered voice—there exists a voyeuristic regime of listening that “wrenches a sexual sound from the body of the other” (21) in order to gratify the male listener with an assurance of their subjective agency.

    In this logic we can discern a trace of the critique of sovereignty advanced by Giorgio Agamben, a thinker whom Pettman evokes on more than one occasion and who, like Pettman, takes his inspiration from the deconstructive logic of exappropriation. Deconstructive essays such as Jacques Derrida’s “Tympan,” for example, suggest that philosophical listening does not simply exclude its outside but seeks to master it and make it its own. But Agamben’s point—as Pettman acknowledges in a note referencing the book Echolalias by Agamben’s translator Daniel Heller-Roazen (100n17)—is that what appears to be outside language is in fact its condition of possibility.[3] As Agamben argues in Language and Death, meaningful human speech can only emerge on condition that the inarticulate animal cry withdraws. Philosophy, though, has traditionally forgotten precisely this withdrawal that makes language possible (what Derrida calls the withdrawal of the withdrawal) and has imagined in its place in its place a bodily presence that appears to lie beyond the bounds of the linguistic. Agamben on the contrary argues that the apparently non-linguistic is nothing other the pure possibility of language that goes unheard in every act of speaking.[4]

    That much of this theory remains in the background leaves Pettman free to write engagingly without getting mired in thorny philosophical debates. Keeping the sustained theorizing largely underground lets Pettman’s prose sparkle. Provocative ideas flow with one intriguing example after another, but this is one of the moments when I would have welcomed a more rigorous corps-à-corps confrontation with Agamben’s theory of Voice. Agamben has a lot to say about what happens when the disavowed condition of possibility begins to circulate in an autonomous sphere—something he specifically connects to analyses of the glorious body, of commodification, and of pornography. Agamben’s commodified body is detached in the pure spectacle from its sacralization, its ineffability and its legally and culturally authorized uses and hence appears as a pure potentiality for new uses. How could Pettman develop Agamben’s reflections on pornography that have always focused on the visual, shifting the focus from visibility to audibility? And how would he situate his own arguments in relation to Agamben’s efforts to dislocate the aporias of metaphysics? When at the beginning of the book, Pettman recalls the prenatal experience of sound, how does this compare with Agamben’s notion of infancy (referenced only in passing at 108n5)? There is little discussion—with the possible exception of Hedy Lamarr’s silent on-screen orgasm—of voices that hold their capacity to sound in reserve.

    Pettman turns in the third chapter to the animal voice. In a chapter indebted to the late Derrida’s ideas on animality, the highlight is a scene with a cockatoo that Pettman contends “deconstructs the cherished metaphysics of (humanist) presence, far more economically and effectively than Derrida does in his writings” (62). The cockatoo was adopted by new owners after a bitter divorce but continues to reenact the no doubt traumatizing arguments it was forced to witness in its previous life with an invective of curse words hurled out with a bitter tone and even the aggravated body language of rejection and resentment. This scene illustrates the difficulty of assigning an owner to the voice: while it is on one level the bird’s voice, audible and present in the room, it also brings to life vividly the original arguing couple. This cockatoo, like the parrot that betrays its owner by reproducing the salacious sounds of the porn he secretly watches, reveals that it is not just imitative animals who are ventriloquized, but we humans too, especially “when we are in the ecstatic, agonistic throes of jouissance or fury.”

    From this Pettman draws the conclusion—albeit one that is hardly new—that there is no simple hierarchy of human over animal, for humans can readily be “reduced” to the “animalistic” under the pressure of certain circumstances. The more thoroughgoing Derridean point that this scene makes—one that Pettman hints at without saying it explicitly—is not only that the human-animal opposition may be deconstructed but that this moreover hinges on a more radical deconstruction of the proper tout court. There is no proper human voice not because humans sometimes cry out in animal voices or because animals sometimes seem to speak to one another. Rather, it is impossible to decide between the two because there is no voice that belongs to any of us, whether human or animal.

    Against a tradition that reserves meaningful speaking and listening as a uniquely human privilege, Pettman thus calls in the final chapter for us to lend our ears to all the voices of the earth, to the vox mundi in which all manner of creatures, entities, and phenomena are present to us. In this Pettman reveals that his concerns are not simply ecological or political but are also properly philosophical, even if he is sometimes coy about asserting this ambition. In other words, Pettman is interested in how Being is present to us as a voice—how it exists for us as we listen to those voices. To this extent, Sonic Intimacy is, despite the framing it often adopts, not chiefly about issues of technology, ecology, or desire. Rather, these themes become occasions to pursue an unashamedly philosophical project: that is, the deconstruction of the metaphysics of voice. To this extent, Pettman’s continues a sequence that extends from Heidegger through French deconstruction: philosophy as listening to Being.

    The parenthetical description in the subtitle “Or, How to Listen to the World,” reveals that there is one philosophical voice in particular that commands Pettman’s attention, even if it is not given the sustained hearing that one might expect. It is Jean-Luc Nancy who tells us, in the face of a rampant globalization that renders the world uninhabitable, that, to be a part of a world and not a mere agglomeration of wealth, we must “share a part of its inner resonances.”[5] Only then can the world take place and can we inhabit it. There are tantalizing references to Nancy scattered throughout the text. There’s a brief mention of his conception of ontology as resonant referral to explain the expropriation of the voice (44–45) and later there’s an unacknowledged and undeveloped evocation of Nancy’s phrase “birth to presence” (89).

    Pettman writes frequently of acousmatic voices where the actual sounding is separated from the source, like the cockatoo. It is tempting, therefore, to imagine Nancy as a kind of disavowed ventriloquist, for Sonic Intimacy—deliberately mixing metaphors here to show the contact between resonance-as-spacing and touch—has Nancy’s fingerprints all over it. The Birth to Presence begins precisely with the same question of defining the human that preoccupies Pettman. The epoch of representation, suggests Nancy, originates with human exceptionalism, with the moment at which the human species being acquired its identity by virtue of one defining characteristic or another. “There is, perhaps, no humanity (and, perhaps, no animality)” wonders Nancy, “that does not include representation.”[6] The task is to think the unraveling of this limit, to think “what, in man, passes infinitely beyond man.” So, if Nancy asks what it is in the human that exceeds the bounds of its exceptional determination, Pettman examines how the exceptional exceeds the bounds of its human definition and thus dissolves the exception. For example, if the human is defined by having a voice, there is part of the human that is not exhausted in its vocality, and there is part of vocality that is not exhausted by the category of the human. Voice and humanity do not coincide. These are two faces of a mutual contamination. Humanity is thereby liberated from its phonocentric determination and vocality spills over the edges of the human into animal cries and the sounds produced by plants, inanimate objects, and intangible algorithms—disseminated throughout the univocity of the vox mundi at large.

    Nancy’s terms of “listening,” “world,” and “being” bear distinctly Heideggerian overtones. Pettman dismisses Heidegger’s suggestion that the animal is poor in world and hence poor in hearing. Adopting Agamben’s critique of what he calls the “anthropological machine” and Derrida’s notion of animot, Pettman has elsewhere not hesitated to point out that Agamben himself fails to get beyond the Heideggerian horizon when he retains boredom, for instance, “as a uniquely human curse and/or privilege.”[7] It is precisely the attunement between beings and their environment that Pettman challenges with his notion of intimacy. He suggests that a sense of self—one intimacy with one’s self if you like—is produced “through the vocal back-and-forths with others—and with the environment” (59). Although Pettman here attributes this notion of back-and-forth to Deleuze and Guattari’s formulation of the refrain, it would surely not have escaped his attention that Nancy describes presence as a “coming and going,” a “back and forth”[8]—what he elsewhere calls a “diapason-subject.”[9]

    This leaves one wondering about the nature of the back-and-forths between Pettman and deconstruction. Does Nancy provide the tools to think about the voice beyond the horizon of anthropogenesis, or are the examples of post-human and non-human voices ways to realize the full implications of Nancy’s deconstruction of sonic presence? One challenge for the reader is that Pettman tends to marginalize precisely those thinkers with whom he is most intimate. He spills more ink, for instance, critiquing Adriana Cavarero than engaging with Derrida. A discussion of the concept of intimacy comes only in the conclusion and many of Pettman’s back-and-forths with deconstruction are reserved to endnotes. One thing that the book could define more clearly is the extent to which the deconstructions of phonocentrism and logocentrism are mutually implicated. In the main body of the text, Pettman suggests that voice is the foundation of logocentrism and in the notes he specifies more precisely that “phōnē is the necessary but not sufficient condition for logos.” Citing Derrida’s claim that phonocentrism appears to be universal, while logocentrism is not, he argues that “the trick is foreground the multitude of voices, without being ‘phonocentric’” (108n8), by which Pettman seems to mean without positing the voice as transcendental.

    There are two questions that remain. First, from the perspective of grammatology: why retain vocality at all even in its plurality? Derrida’s famous attack on Husserl targets the false notion that one is simultaneously present to oneself in hearing-oneself-speak. Already in Husserl the account of temporalization reveals that the supposed unity of the “now” is in fact divided from it—that is, is always already spacing. This is why Pettman insists, against Cavarero, on the significance of time-shifted contexts, in which presence is dispersed. The question remains, though: why continue to speak of a voice if one is thinking of something closely approximating Nancy’s resonant referral? One possible answer is that these voices stripped of logos and bodily presence, represent a pure intention to signify—something close to Agamben’s notion of Voice as the potentiality for language. As Nancy develops the idea that listening-as-resonance is the condition of possibility for sense, he cites a passage from Agamben in which he thinks of Voice as the rustling of animals in their retreat. It would be fascinating to see Pettman engage with this citation in order to specify more precisely the relation between voice and listening. For Pettmann, this relation is defined by the concept of intimacy, according to which a voice is what strives to make itself known to us, which calls us to pay attention to it, summons our listening and invites us to approach its “potentially enlightening alterity” (83). While Pettman is eager to distance himself from neo-Heideggerianism, what prevents this seductive allude from repeating the logic of the withdrawal of Being when the deictic voix-là that he coins, like Agamben’s Voice-as-shifter, consists in its own vanishing act (58)?

    The other point to make is one that could also be leveled at deconstruction: is dispersal and dissemination really an effective way to relinquish the transcendental? Pettman is clearly with Derrida on this point, but Catherine Malabou has made a convincing argument that Derrida’s attraction to a Genetian dissemination of aurality as a means to topple the Hegelian tower of Klang is just another attempt to avoid the economy of the transcendental without abandoning it.[10] The problem with the transcendental voice, as Pettman recognizes, is that it always presupposes another excluded voice. The category of human voice presupposes the other voice of machine and animal, but, even within the category of the human, the voice is divided into noise and speech, masculine and feminine, and so forth, always partitioning itself. In the economy of the transcendental, the voice becomes a fetish—which, in Derrida’s definition, can both be detached from a chain of voices to become the privileged one and also substitute for any other one in the chain.

    One can escape the contradiction by incorporating the externalized fetish into the system (the Hegelian metaphysical solution) or, as Malabou points out, you can deflate the phallus by bringing down everything around it so that nothing stands taller than anything else (the Derridean option). Pettman, for his part, challenges the privileged position of the voice and instead indulges in the substitution of one voice for another, a gradual slippage from one chapter to the next. The issue facing deconstruction applies here too, though: how to end the infinite regress of voices? In the end Pettman seems to settle for a voice of the world that is without beginning or end and that refuses to be subordinated to any totalizing project. The world is a space in which one is always listening out for another voice. One moment one hears it, the next one doesn’t.

    The form and style of Pettman’s book capture the character of this roving ear, always pricking up with the possibility of another intriguing example. Pettman is a very engaging writer, and the way he traverses contexts and theoretical horizons is thrilling. Sonic Intimacy slides from one voice into another, slipping out of one body into another, all the more easily because it wears its weighty themes very lightly. Philosophy, then, becomes less an instrument by which to prosecute an argument than a playful seduction designed to lure our ears from one idea to the next. Pettman’s writing is perhaps at its most exciting when it ignores expectations to pin down the voices of interlocutors and instead revels in throwing the voice, in making it seem as if it emanates from somewhere else. Pettman himself, whose body of writing gives the impression of an insatiable curiosity, is no doubt already chasing down other voices and other worlds. I urge readers, though, to let their ear linger a little longer over this intriguing little book that promises to help us discern voices where we least expect to hear them.

    Naomi Waltham-Smith is Assistant Professor of Music at the University of Pennsylvania. Her work sits at the intersection of music, sound studies, and continental philosophy. She is author of Music and Belonging Between Revolution and Restoration published by Oxford University Press, and is currently writing a book entitled The Sound of Biopolitics.

    Notes

    [1]    Dominic Pettman, Creaturely Love: How Desire Makes Us More and Less Than Human (Minneapolis: University of Minnesota Press, 2017).

    [2]    Dominic Pettman, Human Error: Species-Being and Media Machines (Minneapolis: University of Minnesota Press, 2011).

    [3]    Daniel Heller-Roazen, Echolalias: On the Forgetting of Language (New York: Zone, 2005).

    [4]    Agamben, Language and Death: The Place of Negativity, trans. Karen E. Pinkus (Minneapolis: University of Minnesota Press, 1991).

    [5]    Jean-Luc Nancy, The Creation of the World, Or, Globalization, trans.  François Raffoul and David Pettigrew (Albany, NY: State University of New York Press, 2007), 42.

    [6]    Jean-Luc Nancy, The Birth to Presence, trans. Brian Holmes et al., (Stanford: Stanford University Press, 1993), 1

    [7]    Dominic Pettman, Human Error, 237n71.

    [8]    Nancy, The Birth to Presence, 5.

    [9]    Jean-Luc Nancy, Listening, trans. Charlotte Mandell (New York: Fordham University Press, 2007), 16.

    [10]   Catherine Malabou, “Philosophy in Erection,” Paragraph 39, no. 2 (2016): 238–48.

  • Ben Murphy – The Universes of Speculative Realism: A Review of Steven Shaviro’s The Universe of Things: On Speculative Realism

    Ben Murphy – The Universes of Speculative Realism: A Review of Steven Shaviro’s The Universe of Things: On Speculative Realism

    Steven Shaviro’s The Universe of Things: On Speculative Realism (2014)

    Reviewed by Ben Murphy

    Steven Shaviro begins The Universe of Things (2014) promising a “new look” at Alfred North Whitehead “in light of” speculative realism. The terms of this preface ought to be reversed though, since what follows Shaviro’s introduction is actually a “new look” at speculative realism “in light of” some Whiteheadean ideas. This distinction is important: readers should not seek out The Universe of Things for an introduction to Whitehead qua Whitehead or even a “new look” at Whitehead vis-à-vis current issues of cultural and critical analysis. (Indeed, better options along these lines include, respectively, Shaviro’s own earlier book, Without Criteria (2009), and the more recent University of Minnesota Press collection The Lure of Whitehead (2014).) Universe, on the other hand, is better described as an attempt to map the cumulative geography of speculative realism, a philosophical movement which Shaviro stresses should be referred to in the plural: speculative realisms. Speculative realisms (and its sibling endeavors like object oriented ontology and new materialism) are perpetually in search of heterodox traditions and forgotten figures—philosophical antecedents sought for foundational credence and inspiration. And in this sense Shaviro’s incorporation of Whitehead is the latest in a lengthening line: Graham Harman recuperates a certain version of Heidegger, Jane Bennett returns to Spinoza and Bergson (among others), and, more far afield still, Ian Hamilton Grant champions Schelling’s Naturphilosophie. But if these and other thinkers raid the archive to consolidate new and distinct philosophical templates, Shaviro’s survey is decidedly more evaluative than constructive. Working Whitehead into the cracks of speculative realism, Shaviro widens that movement’s internal fractures in order to expose, and at most nuance—rather than overturn, reverse, or revamp—its prevailing assumptions.

    Shaviro’s critical take on speculative realism relies on two recurring moves: first, an overarching unification and, second, a subsidiary distinction. First, in the name of unity, Shaviro stresses that speculative realisms hold in common a core desire to step outside what he—following French philosopher Quentin Meillasoux—calls the correlationist circle. As reiterated by Shaviro, the primary target implied by this phrase is Kant’s position that the world is only knowable and approachable through thought. “We” can never grasp an object “in itself” or “for itself” in isolation from its relation to us, the thinking subjects. This insistence means that any account of the world and reality is fundamentally an account of the world and reality as accessed through and by human thought. Speculative realisms are unified in wanting to get beyond this self-reflexive loop. Quentin Meillasoux, Graham Harman, Ray Brassier, and Ian Hamilton Grant (the school’s four founding fathers)—as well as fellow travelers—shed the correlationist straight jacket by theorizing (or, better, speculating) about the real world, the world of the “great outdoors” (another Meillasoux coinage) or, as Eugene Thacker puts it in his “horror of philosophy” series, the world “without us.” (For a very different account which disputes whether “correlationism” refers to a fair or even a meaningful reading of Kant, see David Golumbia’s “’Correlationism’: The Dogma that Never Was,” recently published in bounday 2.) As Shaviro notes, there’s a timeliness to this “anti-correlationist” critique, since casting the philosophical net beyond the circumscribing human mind seems a deadly serious endeavor in the face of impending ecological catastrophe. Still, the warming planet is just the most obvious and palatable hook that initiates what Shaviro calls the “changed climate of thought” (4) recently amenable to speculative realism. And if both new materialism and object oriented ontology are more prone to non- or para-academic environmental and ecological interventions, then speculative realism is more interested in revisiting and recasting the history of philosophy.

    A commitment to outfoxing correlationism unites speculative realism, but Shaviro’s second move—that of division—hinges on pinpointing the particular strategies employed to achieve this revisionary project. Repeatedly in Universe, Shaviro splits speculative realism into two main factions. On the one hand, Meillasoux and Brassier pursue lines of thought that Shaviro calls “eliminativist”: for these admittedly nihilistic thinkers, correlationism is undone by the revelation that thought is “epiphenomenal, illusory, and entirely without efficacy” (73)—that thought doesn’t rightly and necessarily belong anywhere in the universe. For Shaviro, Brassier goes further in approaching the “extinction of thought” than Meillasoux, who saves thought from complete elimination by introducing a deus ex machina according to which thought and life emerge “ex nihilo” and simultaneously from a universe previously devoid of both (76). The contrast to this first faction is found in Harman, Grant, Levi Bryant, and Timothy Morton. Instead of proposing that thought is fundamentally inimical to the universe, this coalition of speculative realism wagers that agency and thought are everywhere. Positing the “sheer ubiquity of thought in the cosmos” (82), this position reaches its apotheosis for Shaviro in a panpsychic vision where all things—animate and otherwise—are sentient (if perhaps not exactly conscious). Shaviro places himself in this second faction only after making a further distinction that separates him from Harman in particular. Whereas Harman, according to Shaviro, stresses the withdrawn nature of objects—withdrawn in the sense that the object must always “recede” from its relations (30)—Shaviro joins Whitehead (and Latour) in making a distinction between epistemological withdrawnness and ontological relations (see 105). Where an object may always hold something in reserve from what is knowable to the perceiving mind (as Harman insists), even this measure of the object that is reserved may be affected and changed by modes of contact that elude knowledge and understanding. Because of “vicarious causation” and “immanent, noncognitive contact” (138, 148) (a mode of contact that Shaviro never satisfactorily distinguishes from more popular usages of the term “affect”), an “occult process of influence” occurs that is “outside” any correlation between “subject and object, or knower and known” (148). The object, then, is not so utterly withdrawn as Harman’s narrowly epistemological account suggests. So between eleminativism and panpsychicism as extremes of the speculative realism spectrum, Shaviro says, we’re faced with a “basic choice” (83).

    Describing correlationism and the various offerings to get beyond it is standard fare for speculative realism. But what Universe lacks in originality it compensates for with breadth of analysis and consistently careful, patient exposition. Shaviro admirably treats a wide swath of speculative realists (plus quite a few philosophical giants from both continental and analytical traditions), and he does so with a tone perpetually modulated for utter clarity. Absent is any of the obfuscating rhetoric or over-the-top claims that one might expect from someone who sets out to correct Kant. In part Shaviro’s achievement stems from his own outsider status. His rich body of academic work—on everything from film studies to music video aesthetics to sci-fi infused accelerationism—as well as the light touch on display here and throughout his superb and eclectic online presence (see: http://www.shaviro.com/) stand him in good stead as a welcome interlocutor and guide. Approaching speculative realism as a kindred but not coincident thinker, he’s able to recapitulate his own coming-to-terms with ideas in a way that translates well to other sympathetic non-initiates.

    Apart from style and tone, though, Shaviro’s approach is also commendable for a self-avowed pragmatism of ideas. In an aside in the first chapter, Shaviro applauds Isabelle Stengers for the insight that “the construction of metaphysical concepts always addresses certain particular, situated needs” (33). “The concepts that a philosopher produces,” Shaviro continues, “depend on the problems to which he or she is responding. Every thinker is motivated by the difficulties that cry out to him or to her, demanding a response” (33). While a fair representation of Shaviro’s own admirably simple and workmanlike prose, these statements also epitomize the generous spirit that urges Universe. Shaviro is careful to explain the fruits and situational benefits of every idea that he treats, perhaps especially those ideas that he wants to challenge—an attractive way of grounding philosophical ideas which, being speculative by definition, sometimes feel quite flighty.

    The discussion of panpsychism that spans chapters four and five is the most exciting and original element of Universe. In part this is because it draws on a body of work in cognitive science and the philosophy of biology that Shaviro knows well and that is fresh fodder for discussions of speculative realism. His discussion in this section also has the added charm of giving itself over to the speculative freedoms afforded to speculative realism itself. As Shaviro recognizes, speculative realism is at its best when it joins with speculative fiction in the common task of “extrapolation” (10). Thus in considering panpsychism we’re teased with the notion that slime molds have thoughts (88). Less bogged down by the minutia of distinctions between this SR thinker and that, Shaviro joins a more diverse group of thinkers to consider, for instance, Thomas Nagel’s question about what it’s like to be a bat. Well aware of the absurdities attendant to a truly panpsychic vision, Shaviro lets speculation carry the day, and it’s a pleasure to follow him through a romp that ties the questions of speculative realism to a longer intellectual tradition of sometimes strange twists and turns.

    Also helpful and fresh for speculative realism—although somewhat hard to square with the rest of this book—is Shaviro’s first chapter, which shows how Emmanuel Levinas helps us appreciate speculative realism even as Whitehead’s “aesthetic” mode of “contrast” departs from Levinas’ “ethical” encounter with the Other. Where for Levinas the encounter trumps self-concern, for Whitehead both self-concern (or “self-enjoyment”) and “concern” for the Other are poles best understand in balancing counterpoint (rather than conflict). Apart from being the most detailed analysis of Whitehead’s thought—and, indeed, his thought as it changed in his long arc of writing—this opening account is valuable for SR in arguing that a commitment to circumventing correlationism need not be an ethical project in the traditional sense. In other words, in Shaviro’s reading of Whitehead, a philosophy geared towards the object world “without us” isn’t premised on care. The problem here and elsewhere in Universe, though, is the fuzzy usage of the term “aesthetic.” As I’ve suggested, chapter one deploys this term opposite Levinasian ethics in a frustratingly negative mode of definition: aesthetics is said to be what is not ethics. While gaining some clarification in the volume’s titular chapter (see 52-54), the aesthetic remains unclear even when given new treatment in a discussion of Kant that occupies the last ten pages of the book. Here “aesthetic” is set against knowledge (or epistemology) rather than ethics, and, as my discussion of Shaviro’s disagreement with Harman suggests, “aesthetic” comes to mean something like noncognitive contact, or “affect.” If these disparate senses of the “aesthetic” are related or even mutually inclusive, Shaviro doesn’t do enough to show how.

    For all its merits, Universe suffers heavily from being stuck between monograph and essay collection. One searches in vain for the absent promise that the book’s chapters can be read collectively or in isolation, approached in order or at random. Such a promise, at least, would admit that the chapters don’t serially build to anything in particular. Lacking this or any other clues from Shaviro, though, we’re faced with seven relatively short offerings that loop back on one another with frustratingly little meta-commentary. Much of the mapping of speculative realism as I’ve described it above via unification and division, for instance, appears essentially verbatim in chapters two, six, and seven. The treatment of Harman—both agreement and disagreement—in particular makes continual reappearance. The same could be said of the discussion of panpsychism, which is interesting the first and perhaps even second time but quickly turns suspect as it is recycled through chapters three, four, and five with only the trimmings changed. The mere fact that bits of argument can appear at the beginning and end of the book in essentially the same form (and with Shaviro seemingly unaware of such repetitions) leaves the reader wondering about the value of a journey that feels constrained to a treadmill. A more cynical reader might look to, and find answer in the book’s editorial meta-data, which reveals that the first three chapters are previously published. Insofar as Universe excels at any one thing, then, it may be at academic entrepreneurialism—a feat of (re)publishing in which a triplet of core essays are surrounded with the sort of rhetorical packing peanuts which actually detract from ideas that would be more forceful as standalone articles. The reader already deep inside the sweep of SR may find plenty in this extended cut edition, but those more casually interested will be better served to read independently (as interests dictate) “Self-Enjoyment and Concern” (on Whitehead, Levinas, and SR), “The Actual Volcano” (Shaviro’s primary disagreement with Harman), and “The Universe of Things” (a broad strokes and bouncy introduction to the promises and riddles of SR, new materialism, and object ontology). Each has gems of insight owed to Shaviro’s exhaustive research, and reading them apart from one another—perhaps even in their original contexts—would lessen the rather tiresome burden of trying to figure out how they all fit together.

    Ben Murphy is a Ph.D. student at the University of North Carolina at Chapel Hill. He works on 19th and 20th century American literature, the history and philosophy of science, and critical theory. His essay on James Dickey’s Deliverance and film adaptation is forthcoming from Mississippi Quarterly (2017), and you can also find his writing at ETHOS: A Digital Review of Arts, Humanities, and Public Ethics and The Carolina Quarterly. Website: http://englishcomplit.unc.edu/people/ben-murphy

  • Devin Zane Shaw — Disagreement and Recognition between Rancière and Honneth

    Devin Zane Shaw — Disagreement and Recognition between Rancière and Honneth

    by Devin Zane Shaw

    In an interview from 2012, Jacques Rancière states in response to a question about the role of dialogue in philosophy:

    I don’t believe in the virtue of dialogue in the form of: here’s a thinker, here’s another thinker, they’re going to debate amongst themselves and that’s going to produce something. My idea is that it’s always books that enter into dialogue and not people….Dialogue is never, for me, what it appears to be, which is something like the lightning flash of an encounter, a live exchange.[i]

    We should, then, approach the recent Recognition or Disagreement: A Critical Encounter on the Politics of Freedom, Equality, and Identity (Columbia University Press, 2016), with a similarly circumspect attitude.

    The core of the book, edited by Katia Genel and Jean-Philippe Deranty, is the debate between Rancière and Axel Honneth that took place at the Institute for Social Research in Frankfurt in June 2009, but it also includes a supplementary text by each author and an essay from each editor. Given that the editors’ essays comprise, at eighty pages, forty-five percent of the text, one should be particularly attentive to the ways in which their interventions shape the reception of the debate that was the book’s occasion. Against the editors, I want to argue here that this debate demonstrates the incompatibility of Honneth’s and Rancière’s respective projects. Moreover, Rancière’s work cannot be reconceptualized in the terms of Honneth’s liberal iteration of critical theory without sacrificing precisely those parts of his thought that are the most inventive, interesting, and politically and intellectually subversive.

    My differences with Genel and Deranty can best be summarized through our respective interpretations of Rancière’s claim that, in his critique of Honneth, he has reconstructed a “‘’ [sic] conception of the theory of recognition” (95). In my view, Rancière critically appropriates the terms of “recognition” to show what it would require to become a theory of dissensus and disagreement. Deranty outlines what he takes to be Rancière’s concern with a theory of recognition that ranges from Althusser’s Lesson to Disagreement in order to demonstrate an “in-principle agreement” between Rancière and Honneth (37). First, he argues that many of the examples from Disagreement are based on historical research that Rancière conducted in the 1970s. Then Deranty adduces passages that mention recognition, such as Alain Faure and Rancière’s “Introduction” to La parole ouvrière (1976), where they refer to political struggle as “the desire to be recognized which communicates with the refusal to be despised” (quoted on 38). He also cites an early interpretation of Pierre-Simon Ballanche’s account of the plebeian revolt on Aventine Hill (an episode which also plays a crucial role in the argument of Disagreement) where Rancière writes that the “rebellion was characterized by the fact that it recognized itself as a speaking subject and gave itself a name.”[ii] Rancière continues, though: “Roman patrician power refused to accept that the sounds uttered from the mouths of the plebeians were speech, and that the offspring of their unions should be given the name of a lineage.”[iii] This description has little to do with Honneth’s account of recognition, in which individuals recognize their freedom and the freedom of others as mediated by established social institutions. And then Deranty concedes that “Rancière just disagrees with some of the key concepts used by Honneth,” which undermines the verbal parallels that he draws upon to signal their agreement (36, my emphasis). Indeed, their principled dispute about their respective concepts undermines the very possibility of an “in-principle agreement.” Therefore, to evaluate the relationship between Rancière’s egalitarian politics and Honneth’s theory of recognition we cannot rely on verbal parallels; instead, we must address how the concepts of recognition and disagreement play out in relation to a theory of the political subject, the relation between politics and the political, and problems concerning what Rancière calls “the police” and social normativity.

    To address these questions, I will begin with the final essay included in Recognition or Disagreement, Honneth’s “Of the Poverty of Our Liberty: The Greatness and Limits of Hegel’s Doctrine of Ethical Life.” Earlier in the book, Honneth claims that “all kinds of political orders have to give a certain description or legitimation for who is included in the political community,” and, indeed, political philosophy often aims to supply the legitimation for a given society’s norms that decide how and whether individuals and their practices are included or excluded from the political order (115). Hegel, on Honneth’s account, demonstrates the logical and practical coherence of the social objectivity of the various types of individual freedom, that is, how freedom relates, through recognition, to politics, work, and love.

    In the book’s concluding essay, Honneth examines, first, how Hegel reconciles two common, subjective concepts of individual freedom within his account of objective freedom as it is realized in ethical life. Both subjective concepts are abstract sides of modern political freedom. For Hegel, the transition to modernity entails conceptualizing social institutions as “making possible the realization of freedom” (160). In other words, on Hegel’s account, individual freedoms are mediated through institutions—and institutions are mediated and produced through the actualization or realization of individual freedoms. Thus, when Hegel reconciles the two subjective concepts of freedom, which approximate what Isaiah Berlin calls negative and positive freedom, he demonstrates that both fail to incorporate the objectivity of freedom as it is embodied in concrete social institutions. According to the “negative” concept of freedom, an individual is free insofar as they are unhindered by the actions of others. While Hegel incorporates this incomplete concept of freedom within his system as “abstract right,” which ensures state protections of individual life, property, and freedom of contract, he faults negative freedom for lacking a positive determination of what the subject can do, socially, with freedom. According to the “positive” concept of freedom, which Hegel largely derives from Kant, the basis of morality is autonomy, the self-legislating and self-reflexive activity of the subject. While this concept of freedom gives a positive foundation to what morality is, it nonetheless remains subjective, lacking a concrete relationship to social objectivity.

    These negative and positive concepts of freedom are, therefore, in Hegel’s terms, “merely” subjective, while Hegel aims to demonstrate that individual freedom is objective, that is, reflected and recognized within objective social institutions. This concept of objective freedom is not limited merely to how we understand social institutions. To say that freedom is objective delimits an important intersubjective feature of individual freedom. As Honneth points out, Hegel argues that we cannot rely on Kantian models of autonomy in friendship or love, since the self-limitation of my freedom in the experience of friendship or love is not a self-limitation; it is “precisely that the other person is a condition of realizing my own, self-chosen ends” (164). The realization of a given individual’s freedom entails concrete social situations that implicate the freedom of others, and it is because social institutions mediate our relations with others that they have objective reality. Hegel—and by extension, Honneth—maintains that institutions receive normative justification insofar as they reflect and embody the practices of individuals’ freedoms, and that social institutions, in turn, engender the emergence and expansion of individual freedoms.

    Now, one can see why Honneth follows Hegel through the discussion of objective freedom in the doctrine of ethical life: what both the negative and positive subjective concepts of freedom lack is recognition. In our institutions, Honneth suggests, we should be able to recognize not only our own intentions but also the intentions of other subjects. In addition, Hegel identifies three ethical spheres in which each individual’s freedoms are realized in relation to others’: personal relationships, the market economy, and politics. For these reasons, Honneth argues that the “general structure” of Hegel’s doctrine of ethical life, despite some shortcomings, “remains sound even today,” and that this doctrine provides “us with a normative vocabulary that we can use to assess the respective value of the various freedoms we practice” (169; 167). Nonetheless, Honneth also faults Hegel for treating “as sacrosanct” three historically specific institutions as the outcome of the self-realization of objective spirit: the family—“guided by the patriarchal prejudices of his own day”—the capitalist market economy, and constitutional monarchy (171). While Hegel did not explicitly address the possibility that these institutions could be transformed to “make them more amenable to the basic demand for relations of reciprocity among equals,” Honneth contends that Hegel’s account of morality hints toward how political practice can revise social norms and reorganize social institutions to make them more democratic (172). According to Honneth’s revision of Hegel, the inclusion of liberal rights and the possibility for “moral self-positioning” allows for individuals to engage in “morally articulated protest” (174). Thus Honneth allows for a continued moral progress within societies and social institutions to a degree that was not envisioned by Hegel.

    *

    Despite his Hegelian framework, and despite his debts to the Frankfurt School, Honneth’s project shares some of the central concerns of mainstream Anglo-American political philosophy today: the emphasis on processes of justification and establishing conditions of justice in order to evaluate institutional and normative frameworks. By contrast, Rancière’s political thought shares neither the methods nor goals of mainstream political philosophy. Todd May has already explored in detail the differences between Rancière and mainstream political philosophy (including Rawls, Nozick, Amartya Sen and Iris Marion Young). In May’s account, these political philosophers rely, whether they are proponents or critics of distributive theories of justice, on a concept of “passive equality”: “the creation, preservation, or protection of equality by governmental institutions.”[iv] Rancière, though, makes the stronger polemical claim that political philosophy embeds itself in, and offers justification for, regimes of inequality that he calls “the police” or “policing.” One of the most striking features of Rancière’s work is his claim that what we typically call politics, even in its most democratic forms (voting, deliberation, governance, and popular legitimation), is policing. In Disagreement, Rancière defines the police as:

    first an order of bodies that defines the allocation of ways of doing, ways of being, and ways of saying, and sees that those bodies are assigned by name to a particular place and task; it is an order of the visible and the sayable that sees that a particular activity is visible and that another is not, that this speech is understood as discourse and another as noise.[v]

    Since this definition of the police sounds very close to the way that Rancière often glosses his concept of “the distribution of the sensible,”[vi] we should specify that policing produces and reproduces relations of inequality, the stratification of roles within a given distribution of the sensible that partition individuals and groups according to inclusion and exclusion, such as those whose task it is to rule and those whose task it is to obey. Moreover, on Rancière’s account, politics—in May’s terms, “active equality”—is a dynamic of collective engagement and revolt that aims to subvert and resist the stratification and coercion of policing and social institutions. Given that Honneth’s account of recognition emphasizes how social institutions mediate and engender individual freedoms, it then follows that in Rancière’s terms Honneth’s theory of recognition would not be an account of politics as much as it is an account—though a progressive one at that—of policing.

    And yet, in “Critical Questions on the Theory of Recognition,” his critique of Honneth (and Chapter Three of the book), Rancière does not use the terms “police” or “policing.” Instead, he begins with the conditional hypothesis that his differences with Honneth are best articulated by treating their respective approaches as competing theories of recognition. At the outset, however, he signals his critical intent by suggesting that “the term ‘recognition’ might also emphasize a relationship between already existing entities,” these entities being individuals and established social institutions (83). When, then, Rancière concludes that he’s sketched, through his critique of Honneth, his own theory of recognition, he’s appropriated the language of critical theory to articulate a politics of dissensus and disagreement.

    Rancière pursues this hypothesis—that he and Honneth are outlining competing theories of recognition—in order to locate their central points of disagreement. In Disagreement, Rancière defines disagreement (la mésentente) as a specific kind of political challenge to a given order of policing, “a determined kind of speech situation in which one of the interlocutors at once understands [entend] and does not understand [entend] what the other is saying.”[vii] In French, the term la mésentente plays on different connotations of the verb entendre, between “to hear” and “to understand.” On Rancière’s account, the politics of disagreement emerges when the marginalized or oppressed (what he calls “the part with no part”) within a given social order challenge the ways in which society is policed, and often these challenges are phrased in terms that have readily accepted meanings within society. However, politically contentious terms, such as equality, rights, or justice are given inventive new meanings that challenge the normative frameworks of a given regime of policing; the part with no part who are contesting injustice and the police can “hear” the same demands but “understand” entirely different things. Many political theorists lament this ambiguity and aim to define it away. However, Rancière argues that the ambiguity of our contentious terms and ideals makes dissensus possible. That is, this ambiguity makes it possible to identify how these politically contentious terms circulate between policing and politics, how they come to articulate and combat inequality and coercion. For example: justice, for some, means due process and equal consideration before the law, while justice for movements such as Black Lives Matter opens on to both a broad indictment of how so-called due process legitimates injustice against African-Americans who are victims of police violence, and a broader vision of transformative social justice.

    In “Critical Questions on the Theory of Recognition,” Rancière uses disagreement in a broader, dialogical sense rather than its specific, political sense. He argues that dialogue—to be truly dialogical—must be an “act of communication [which] is already an act of translation, located on a terrain that we don’t master” (84). Dialogue always involves translation, distortion, but also invention; in terms of philosophy, it means that both interlocutors must think outside of their usual terminology: distortion remains “at the heart of any mutual dialogue, at the heart of the form of universality on which dialogue relies” (84). But Rancière also suggests that dialogue, in its more specific, political sense, requires acknowledging the “asymmetry in positions” between interlocutors. This claim summarizes his differences with Habermas, which he had previously outlined in Disagreement: acknowledging how asymmetry and power distort the ideals of political dialogue entails, in Rancière’s account, a stringent form of universalism that demands philosophers to confront not just institutional barriers to democratic deliberation, but also how the processes of deliberation function to exclude certain forms of political speech and action. Thus Rancière’s critical question: to what degree does Honneth’s theory of recognition rely on the presupposition that the demands of political subjects have always already been mediated by social institutions?

    To confront this question, Rancière proposes three working definitions of recognition. Two reflect common usage: on the one hand, recognition means the concurrence of a perception with prior knowledge, as when we recognize a friend, location, or information; on the other hand, recognition in the moral sense designates how we recognize other individuals as autonomous beings like ourselves. In both cases, Rancière notes, “re-cognition” functions as an act of confirmation. He then hypothesizes that recognition could also be conceptualized in the terms of what he calls a distribution of the sensible. Recognition, then, “focuses on the configuration of the field in which things, persons, situations, and arguments can be identified” (85). In this sense, recognition comes prior to any act of confirmation—and the critique of recognition entails disagreement over the conditions in which persons, things, or situations are understood as such.

    We could ask, for instance, how is it that a given regime of policing frames some enunciations as political demands against injustice and others as merely subjective complaints or even noise? And we could use an analysis of this situation to attack the broader norms that legitimate this distribution of speech and noise.  While Rancière acknowledges that Honneth’s account of recognition “echoes” his own polemical account, he raises a crucial question: to what degree does Honneth’s account rely on the two connotations of the common usage, presuming a stable distribution of the sensible or normative framework that relies on an “identitarian conception of the subject” that conflicts with a “conception of social relations as mutual” or dynamically or socially constructed (85)?

    First, Rancière contends that Honneth embraces an “anthropological-psychological” concept of the subject that is heavily indebted to a Hegelian “juridical definition of the person” (87). Thus Honneth’s account of the subject’s struggle for recognition emphasizes the affirmation of self-identity and self-integrity within the intersubjective structure of recognition. In other words, it’s the same integral individual subject who is seeks recognition within a multiplicity of situations related to love, work, or politics. Then, Rancière argues that this juridical model of the integral identity of the subject conflicts with its claim to articulating intersubjective social agency—a point encapsulated in Honneth’s summary of love and recognition in the book: “in friendship and love my experience is precisely that the other person is a condition of my realizing my own, self-chosen ends” (164). To say that love involves two individuals realizing their respective ends and interests through another is overly juridical. To Honneth, Rancière counterposes love as it is found in À la recherche du temps perdu, where Proust describes love as a dynamic and aesthetic construction of an other. Rancière writes:

    What appears at the beginning is the confused apparition of a multiplicity, an impersonal patch on a beach. Slowly the patch appears as a group of young girls, but is still a kind of impersonal patch. There are many metamorphoses in that patch, in the multiplicity of young girls, through to the moment when the narrative personifies this impersonal multiplicity, gives it the face of one person, the object of love, Albertine. (88)

    Rancière offers this counternarrative to show how our theoretical frameworks delimit the possibilities of social agency that we are able to recognize—a criticism that Honneth subsequently accepts.

    Rancière’s attention to this point can perhaps explain how Rancière’s terminology can be alternately powerful and abstract. When he opposes the politics of equality to policing, it readily calls to mind clashes between protestors and cops, though politics cannot be reduced to these terms. However, when he defines those subjects who confront the established order as the part with no part, this definition is far more abstract than saying marginalized and oppressed. But Rancière relies on this level of abstraction in order to avoid delimiting conditions of political agency that could delimit who this part is because it could exclude groups who have yet to emerge and who we cannot foresee.

    In general, for Rancière, political subjects are neither self-identical nor self-integral. Instead, political subjects emerge through a dynamic of what he calls disidentification, the rejection of the roles, places, and tasks assigned to bodies within a given regime of policing. We could interpret Proust’s description of love, then, as a metaphor for the dynamic of political subjectivation: political subjects emerge first as a multiplicity, at first an impersonal patch in the social field, until it takes shape through the invention of a name—for instance, #blacklivesmatter or #NoDAPL—for a collective disruption of or rebellion against the police order. Given that all regimes of policing are instantiations of social inequality and coercion, politics is, for Rancière, by definition egalitarian. It is equality, he argues, that leads to a much more exacting concept of universality than an account of politics that neglects the asymmetry between the political subjects who exist by virtue of contesting the social order and the established order of policing. Politics enacts the affirmation of “an equal capacity to discuss common affairs”; in other words, politics enacts the intellectual and political equality of anybody and everybody (93).

    The task of political thought is to ascertain how politics involves a “polemical configuration of the universal” (94). The Black Lives Matter movement began with a call for justice for Mike Brown in Ferguson, but, according to Keeanga-Yamahtta Taylor, its next stage involves both “engaging with the social forces that have the capacity to shut down sectors of work or production until our demands to stop police terrorism are met” and movement building through solidarity, which addresses how, while African-Americans “suffer most from the blunt force trauma of the American criminal justice system,” the broader normative framework of “law-and-order politics” functions to oppress the poor in general.[viii] From a standpoint informed by Rancière, the goal of political thought would be to identify the movements and practices that drive “the process of spreading the power of equality” in the here and now, to identify how specific movements involve a polemical force of universality to subvert and combat the normative frameworks of a given police order (94). Far from endorsing a theory of recognition, Rancière has redefined recognition as a politics of dissensus and disagreement.

    *

    Thus we have good reason to doubt Deranty’s claim of an in-principle agreement between Rancière and Honneth. Indeed, the editors and I reach very different conclusions regarding the significance of this debate because they accept Honneth’s theoretical framework to interpret it, while I refuse to subsume Rancière’s concepts under Honneth’s. The point here, though, is not establishing who has read Rancière or Honneth correctly, but to examine how these interpretations delimit what each thinker believes is politically possible and feasible.

    Our first difference concerns the supposed common ground shared by Rancière and Honneth. Though Rancière explicitly chooses to oppose “politics” (la politique), rather than “the political” (le politique) to “the police”, Honneth and the editors equivocate between “politics” and “the political.” However, the terms, especially in French philosophy, are distinct—which means Rancière has made a deliberate conceptual choice.[ix] Politics, on his account, designates a dynamic activity, while “the political” carries the connotation of an original, fundamental political sphere upon which policing has supervened. For Honneth, then, when Rancière discusses equality, he’s describing either an “original definition of the political community” (115) or a political anthropology in which human beings “are constituted by a wish or a desire to be equal to all others,” and this “egalitarian desire…brings about the exceptional moment of politics” (99). In their “Critical Discussion” included in the book, Rancière rightly rebuts both of these characterizations. He holds that if politics takes place, it does so through an egalitarian praxis opposed to the police. To treat Rancière’s politics as a political anthropology, imputing particular desires or motives to political subjects, implies that the debate is about whether human beings are motivated by either a desire for recognition or for equality. We could, in that case, resolve the debate with a political anthropology of desire.

    If this is not enough reason to reject Honneth’s way of framing the debate, he also characterizes recognition and disagreement as two complementary forms of struggle with different scopes—but this categorization carries with it an implicit normative claim that recognition is more practical. He argues that Rancière brusquely reduces “the political,” considered as “a stratified normative order of principles of recognition,” to policing (103). Therefore Rancière interprets this stratified normative order too rigidly, when these norms are given to conflicts over their meaning, that is, subject to reinterpretation and revision. For Honneth, the revisability of the normative order allows us to conceive of two types of political intervention: an internal struggle for recognition and an external struggle for recognition. In Honneth’s terms, Rancière focuses exclusively on the external struggle for recognition, which, while it combats the “political order as such,” ignores the “reformist” ambitions of the internal struggle for recognition that aims to reinterpret existing normative principles to make social institutions and their normative frameworks more democratic and inclusive.

    But Honneth’s distinction between the internal and external struggles for recognition is not merely descriptive, but also normative: given, he claims, the difficulties in formulating injustice in revolutionary terms, it’s more important in day-to-day politics to “deal with these small projects of redefinition or of reappropriation of the existing modes of political legitimation” (106). Unlike Honneth, Rancière does not prescribe the scope of political struggle within a given situation, since such a prescription functions to legitimate or delegitimize choices we make about what is to be done. These choices cannot be evaluated outside of the context of political struggle itself. But Honneth’s normative preference is part of his philosophical framework: if the freedom of individuals is engendered and mediated by social institutions and norms, and if self-integrity is one of the primary ends of the theory of recognition, then individuals should aim to reform and reinterpret these institutions and norms incrementally.

    From Rancière’s perspective, even if we grant that political freedom is sometimes engendered by existing social institutions, this does not entail that all parts of society should recognize these institutions as engendering their freedom. Those who are marginalized and oppressed could just as equally recognize how a given institution has functioned to exclude, marginalize, oppress, or immiserate them. The goal of politics for these political subjects need not or should not be—nor should we prescribe their goal to be—the reform of or formal recognition within these institutions that have historically oppressed them. From Rancière’s standpoint, it is right for the part with no part to combat and transform the very normative principles that legitimate and reinforce these institutions of inequality, and to prescribe reform rather than radical normative transvaluation serves to delegitimize the possibility of formulating and enacting broader goals of political struggle.

    Thus while Recognition or Disagreement presents the debate between Rancière and Honneth, it speaks to broader issues about the scope and aims of contemporary political thought. The contrast between Honneth and Rancière ably demonstrates Rancière’s stubborn refusal to engage in the processes of justification valorized by mainstream political theory—indeed, it serves as a stark reminder of how engaging in these problems often, (and in Rancière’s view, always) entails accepting profound social inequalities. However, this book is also important because it shows that if we mainstream Rancière’s work, as Genel and Deranty attempt to do, we lose those parts of his work that are most subversive and inventive—and we are left with only Honneth.

    Devin Zane Shaw teaches philosophy at Carleton University. He is the author of Egalitarian MomentsFrom Descartes to Rancière (Bloomsbury, 2016) and Freedom and Nature in Schelling’s Philosophy of Art (Bloomsbury, 2010).

    Notes

    [i] Jacques Rancière, The Method of Equality: Interviews with Laurent Jeanpierre and Dork Zabunyan, transl. Julie Rose. Malden: Polity, 2016, p. 183.

    [ii] Quoted on 38, but the reference is incomplete. See Rancière, Staging the People: The Proletarian and His Double, transl. David Fernbach. London: Verso, 2011, p. 37.

    [iii] Rancière, Staging the People, 37.

    [iv] Todd May, The Political Thought of Jacques Rancière: Creating Equality. University Park: Pennsylvania State University Press, 2008, p. 3.

    [v] Rancière, Disagreement: Politics and Philosophy, transl. Julie Rose. Minneapolis: University of Minnesota Press, 1999, p. 29.

    [vi] As Rancière defines it in Recognition or Disagreement, a distribution of the sensible is “a relation between occupations and equipments, between being in a specific space and time, performing specific activities, and being endowed with capacities of seeing, saying, and doing that ‘fit’ those activities. A distribution of the sensible is a set of relations between sense and sense, that is, between a form of sensory experience and an interpretation that makes sense of it. It is a matrix that defines a whole organization of the visible, the sayable, and the thinkable” (136).

    [vii] Disagreement, p. x.

    [viii] Keeanga-Yamahtta Taylor, From #BlackLivesMatter to Black Liberation. Chicago: Haymarket Books, 2016, pp. 217, 211.

    [ix] See Samuel A. Chambers, The Lessons of Rancière. Oxford: Oxford University Press, 2013, pp. 50–57.

     

  • Quinn DuPont – Ubiquitous Computing, Intermittent Critique

    Quinn DuPont – Ubiquitous Computing, Intermittent Critique

    a review of Ulrik Ekman, Jay David Bolter, Lily Díaz, Morten Søndergaard, and Maria Engberg, eds., Ubiquitous Computing, Complexity, and Culture (Routledge 2016)

    by Quinn DuPont

    ~

    It is a truism today that digital technologies are ubiquitous in Western society (and increasingly so for the rest of the globe). With this ubiquity, it seems, comes complexity. This is the gambit of Ubiquitous Computing, Complexity, and Culture (Routledge 2016), a new volume edited by Ulrik Ekman, Jay David Bolter, Lily Díaz, Morten Søndergaard, and Maria Engberg.

    There are of course many ways to approach such a large and important topic: from the study of political economy, technology (sometimes leaning towards technological determinism or instrumentalism), discourse and rhetoric, globalization, or art and media. This collection focuses on art and media. In fact, only a small fraction of the chapters do not deal either entirely or mostly with art, art practices, and artists. Similarly, the volume includes a significant number of interviews with artists (six out of the forty-three chapters and editorial introductions). This focus on art and media is both the volume’s strength, and one of its major weaknesses.

    By focusing on art, Ubiquitous Computing, Complexity, and Culture pushes the bounds of how we might commonly understand contemporary technology practice and development. For example, in their chapter, Dietmar Offenhuber and Orkan Telhan develop a framework for understanding, and potentially deploying, indexical visualizations for complex interfaces. Offenhuber and Telhan use James Turrell’s art installation Meeting as an example of the conceptual shortening of causal distance between object and representation, as a kind of Peircean index, and one such way to think about systems of representation. Another example of theirs, Natalie Jermijenko’s One Trees installation of one hundred cloned trees, strengthens and complicates the idea of the causal index, since the trees are from identical genetic stock, yet develop in natural and different ways. The uniqueness of the fully-grown trees is a literal “visualization” of their different environments, not unlike a seismograph, a characteristic indexical visualization technology. From these examples, Offenhuber and Telhan conclude that indexical visualizations may offer a fruitful “set of constraints” (300) that the information designer might draw on when developing new interfaces that deal with massive complexity. Many other examples and interrogations of art and art practices throughout the chapters offer unexpected and penetrating analysis into facets of ubiquitous and complex technologies.

    James Turrell, Meeting 2016
    MoMA PS1 | James Turrell, Meeting 2016, Photos by Pablo Enriquez

    A persistent challenge with art and media analyses of digital technology and computing, however, is that the familiar and convenient epistemological orientation, and the ready comparisons that result, are often to film, cinema, and theater. Studies reliant on this epistemology tend to make a range of interesting yet ultimately illusory observations, which fail to explain the richness and uniqueness of modern information technologies. In my opinion, there are many important ways that film, cinema, and theater are simply not like modern digital technologies. Such an epistemological orientation is, arguably, a consequence of the history of disciplinary allegiances—symptomatic of digital studies and new media studies originating from screen studies—and a proximate cause of Lev Manovich’s agenda-setting Language of New Media (2001), which relished in the mimetic connections resulting from the historical quirk that the most obvious computing technologies tend to have screens.

    Because of this orientation, some of the chapters fail to critically engage with technologies, events, and practices largely affecting lived society. A very good artwork may go a long way to exposing social and political activities that might otherwise be invisible or known only to specialists, but it is the role of the critic and the academic to concretize these activities, and draw thick connections between art and “conventional” social issues. Concrete specificity, while avoiding reductionist traps, is the key to avoiding what amounts to belated criticism.

    This specificity about social issues might come in the form of engagement with normative aspects of ubiquitous and complex digital technologies. Instead of explaining why surveillance is a feature of modern life (as several chapters do, which is, by now, well-worn academic ground), it might be more useful to ask why consumers and policy-makers alike have turned so quickly to privacy-enhancing technologies as a solution (to be sold by the high-technology industry). In a similar vein, unsexy aspects of wearable technologies (accessibility) now offer potential assistance and perceptual, physical, or cognitive enhancement (as described in Ellis and Goggin’s chapter), alongside unprecedented surveillance and monetization opportunities. Digital infrastructures—both active and failing—now drive a great deal of modern society, but despite their ubiquity, they are hard to see, and therefore, tend not to get much attention. These kinds of banal and invisible—ubiquitous—cases tend not to be captured in the boundary-pushing work of artists, and are underrepresented (though not entirely absent) in the analyses here.

    A number of chapters also trade on old canards, such as worrying about information overload, “junk” data whizzing across the Internet, time “wasted” online, online narcissism, business models based on solely on data collection, and “declining” privacy. To the extent that any of these things are empirically true—when viewed contextually and precisely—is somewhat beside the point if we are not offered new analyses or solutions. Otherwise, these kinds of criticisms run the risk of sounding like old people nostalgically complaining about an imagined world before technological or informational ubiquity and complexity. “Traditional” human values might be an important form of study, but not as the pile-on Left-leaning liberal romanticism prevalent in far too many humanistic inquiries of the digital.

    Another issue is that some of the chapters seem to be oddly antiquated for a book published in 2016. As we all know, the publication of edited collections can often take longer than anyone would like, but for several chapters, the examples, terminology, and references feel unusually dated. These dated chapters do not necessarily have the advantage of critical distance (in the way that properly historical study does), and neither do they capture the pulse of the current situation—they just feel old.

    Before turning to a sample of the truly excellent chapters in this volume, I must pause to make a comment about the book’s physical production. On the back cover, Jussi Parikka calls Ubiquitous Computing, Complexity, and Culture a “massively important volume.” This assessment might have been simplified by just calling it “a massive volume.” Indeed, using some back-of-the-napkin calculations, the 406 dense pages amounts to about 330,000 words. Like cheesecake, sometimes a little bit of something is better than a lot. And, while such a large book might seem like good value, the pragmatics of putting an estimated 330,000 words into a single volume requires considerable care to typesetting and layout, which unfortunately is not the case here. At about 90 characters per line, and 46 lines per page—all set in a single column—the tiny text set on extremely long lines strains even this relatively young reviewer’s eyes and practical comprehension. When trudging through already-dense theory and the obfuscated rhetoric that typically accompanies it (common in this edited collection), the reading experience is often painful. On the positive side, in the middle of the 406 pages of text there are an additional 32 pages of full-color plates, a nice addition and an effective way to highlight the volume’s sympathies in art and media. An extensive index is also included.

    Despite my criticisms of the approach of many of the chapters, the book’s typesetting and layout, and the editors’ decision to attempt to collocate so much material in a single volume, there are a number of outstanding chapters, which more than redeem any other weaknesses.

    Elaborating on a theme from her 2011 book Programmed Visions (MIT), Wendy H.K. Chun describes why memory, and the ability to forget, is an important aspect to Mark Weiser’s original notion of ubiquitous computing (in his 1991 Scientific American article). (Chun also notes that the word “ubiquitous” comes from “Ubiquitarians,” a Lutherans sect who believed Christ was present ‘everywhere at once’ and therefore invisible.) According to Chun’s reading of Weiser, to get to a state of ubiquitous computing, machines must lose their individualized identity or importance. Therefore, unindividuated computers had to remember, by tracking users, so that users could correspondingly forget (about the technology) and “thus think and live” (161). The long history of computer memory, and its rhetorical emergence out of technical “storage” is an essential aspect to the origins of our current technological landscape. Chun notes that prior to the EDVAC machine (and its strategic alignment to cognitive models of computation), storage was a well understood word, which etymologically suggested an orientation to the future (“stores look toward a future”). Memory, on the other hand, contained within it the act of recall and repetition (recall Meno’s slave in Plato’s dialogue). So, when EDVAC embedded memory within the machine, it changed “memory by making memory storage” (162). In doing so, if we wanted to rehabilitate Weiser’s original image, of being able to “think and live,” we would need to refuse the “deadening of the world brought about by memory as storage and realize the fundamentally collective nature of memory and writing” (162).

    Sean Cubitt does an excellent job of exposing the political economy of ubiquitous technologies by focusing on the ways that enclosure and externalization occur in information environments, interrogating the term “information economy.” Cubitt traces the history of enclosures from the alienation of fifteenth-century peasants from their land, the enclosure of skills to produce dead labour in nineteenth-century factories, to the conversion of knowledge into information today, which is subsequently stored in databases and commercialized as intellectual property—alienating individuals from their own knowledge. Accompanying this process are a range of externalizations, predominantly impacting the poor and the indigenous. One of the insightful examples Cubitt offers of this process of externalization is the regulation of radio spectrum in New Zealand, and the subsequent challenge by Maori people who, under the Waitangi Treaty, are entitled to “all forms of commons that pre-existed the European arrival” (218). According to the Maori, radio spectrum is a form of commons, and therefore, the New Zealand government is not permitted to claim exclusive authority to manage the spectrum (as practically all Western governments do). Not content to simply offer critique, Cubitt concludes his chapter with a (very) brief discussion of potential solutions, focusing on the reimagining of peer-to-peer technology by Robert Verzola of the Philippines Green Party. Peer to peer technology, Cubitt tentatively suggests, may help reassert the commons as commonwealth, which might even salvage traditional knowledge from information capitalism.

    Katie Ellis and Gerard Goggin discuss the mechanisms of locative technologies for differently-abled people. Ellis and Goggin conclude that devices like the later-model iPhone (not the first release), and the now-maligned Google Glass offer unique value propositions for those engaged in a spectrum of impairment and “complex disability effects” (274). For people who rely on these devices for day-to-day assistance and wayfinding, these devices are ubiquitous in the sense Weiser originally imagined—disappearing from view and becoming integrated into individual lifeworlds.

    John Johnston ends the volume as strongly as N. Katherine Hayles’s short foreword opened it, describing the dynamics of “information events” in a world of viral media, big data, and, as he elaborates in an extended example, complex and high-speed financial instruments. Johnston describes how events like the 2010 “Flash Crash,” when the Dow fell nearly a thousand points and lost a trillion dollars and rebounded within five minutes, are essentially uncontrollable and unpredictable. This narrative, Johnston points out, has been detailed before, but Johnston twists the narrative and argues that such a financial system, in its totality, may be “fundamentally resistant to stability and controllability” (389). The reason for this fundamental instability and uncontrollability is that the financial market cannot be understood as a systematic, efficient system of exchange events, which just happens to be problematically coded by high-frequency, automated, and limit-driven technologies today. Rather, the financial market is a “series of different layers of coded flows that are differentiated according to their relative power” (390). By understanding financialization as coded flows, of both power and information, we gain new insight into critical technology that is both ubiquitous and complex.

    _____

    Quinn DuPont studies the roles of cryptography, cybersecurity, and code in society, and is an active researcher in digital studies, digital humanities, and media studies. He also writes on Bitcoin, cryptocurrencies, and blockchain technologies, and is currently involved in Canadian SCC/ISO blockchain standardization efforts. He has nearly a decade of industry experience as a Senior Information Specialist at IBM, IT consultant, and usability and experience designer.

    Back to the essay

  • Andrew Martino – Exhuming the Text: Alice Kaplan’s “Looking for the Stranger: Albert Camus and the Life of a Literary Classic”

    Andrew Martino – Exhuming the Text: Alice Kaplan’s “Looking for the Stranger: Albert Camus and the Life of a Literary Classic”

    Alice Kaplan’s Looking for the Stranger: Albert Camus and the Life of a Literary Classic

    Reviewed by Andrew Martino

    Albert Camus never considered himself an existentialist. In fact, Camus never exclusively believed in any school of thought. Camus was the consummate outsider, the one who stood apart from those who subscribed to views that forced those subscribers into a narrow ideology, especially when that ideology mixed with violence, something Camus steadfastly resisted. If we had to place Camus into any category it would be that of the humanist caught in the absurd. Camus believed in life over death (without believing in an afterlife), yet this belief did not keep him from contemplating the question of suicide, the only serious philosophical problem confronting us, as he writes in The Myth of Sisyphus. Camus’ humble beginnings in extreme poverty and illiteracy in his native Algeria  testify to the power of the human spirit in the face of an indifferent world. When he was awarded the Nobel Prize for Literature in 1957 he expressed reservations and claimed that the prize should have gone to André Malraux, an early influence on his writing. Camus also realized that the Nobel would bring a certain celebrity that would complicate his life, perhaps even sabotaging his art. Add to this his “silence” on the Algerian problem and his very public and acrimonious break with Sartre, and Camus becomes a figure trapped in a world where he is increasingly unable to control his own image. Camus is a problematic figure who is claimed by both the Right and the Left, leaving the man and his writing caught in a political vortex. Focusing on the postcolonial aspect of The Stranger, Edward W. Said writes that Camus “is a moral man in an immoral situation.”[i] When Camus died at the age of 46 in a car accident in 1960, he left the world with the image of the charismatic young man, Bogart-like in his coolness, and still with the promise of great things to come. But a saint he was not. His numerous affairs and constant womanizing, his reluctance to act or speak out against French imperialism in Algeria, his disillusionment with and expulsion from the Communist Party, render him more human than academics might be comfortable with. Camus’ life was full of contradictions, full of silences. Yet, it was precisely from these contradictions and silences that Camus produced one of the most important and widely read books of the twentieth century.

     Looking back over the seven decades since the publication of The Stranger, Camus’ reluctance to situate (in the Sartrean sense of the term) himself in the bubble of existentialism, a bubble in which The Stranger and his relationship with Sartre placed him, the novel blazed a path that opened up fields where the absurd might be articulated, contemplated, and confronted from the inside (the modernist bent) rather than from above and beyond, as the canonical novels of the nineteenth century may have done. In her essay “French Existentialism,” Hannah Arendt briefly examines Sartre and Camus’ influence on the “new” movement where novels carry the weight of philosophy. Throughout that essay she also comments on Camus’ reluctance to be labeled an existentialist. “Camus has probably protested against being called an Existentialist because for him the absurdity does not lie in man as such or in the world as such but only in their being thrown together.”[ii] Here we have what is perhaps the most concise and articulate formulation of absurdist philosophy to date. Camus’ definition of absurdity, painstakingly mapped out in Caligula, The Stranger, and The Myth of Sisyphus, is not quite existentialism, but does contain existentialist DNA, especially Kierkegaardian and Dostoevskian (two of Camus’ patron saints) DNA. As Camus remarks in The Myth of Sisyphus: “I can therefore say that the Absurd is not in man (if such a metaphor could have a meaning) nor in the world, but in their presence together.”[iii] Camus’ definition of the absurd is also the epistemological curve in the road separating him from Sartre’s thinking. If Sartre’s philosophy can be distilled into his phrase “Hell is other people,” then Camus’ is a philosophy of the absurd dependent upon relationships among people. On the other hand, Camus’ articulation of the absurd, as we’ve seen above, resides in the relationship of humans with their world.

    Together, Sartre and Camus blazed a path where philosophy and art, in this case literature, met, thereby ushering in a new form of the novel, one that would examine existence from a philosophical perspective while making use of a form in which to mold these philosophical perspectives. What emerges from this is a hybrid. According to Randall Collins, “What was identified was a tradition of literary-philosophical hybrids. Sartre and Camus were key formulators of the canon, and themselves archetypes of the career overlap between academic networks and the writers’ market. The phenomenon of existentialism in the 1940s and 1950s added another layer to this overlap.”[iv] But this hybridization was more than a heady cerebral new movement in fiction; this hybrid constituted a new way of thinking about the world, a world that emerged primarily from a particular network of intellectuals at a particular time in Paris. Sartre and Camus are on the crest of this wave of existentialism and their thinking would go on to change the world.

    Alice Kaplan’s extraordinary new book Looking for the Stranger: Albert Camus and the Life of a Literary Classic, is a careful and meticulously researched examination of Camus’ 1942 novel. Kaplan is one of the leading scholars of twentieth century French culture and history. She is currently the John M. Masser Professor of French at Yale University where she also received her Ph.D. in French in 1981. She has published seven books, including: French Lessons: A Memoir (1993), The Collaborator: The Trial and Execution of Robert Brasillach (2000), and Dreaming in French: The Paris Years of Jacqueline Bouvier Kennedy, Susan Sontag, and Angela Davis (2012). In 2013 Kaplan edited and provided the introduction to The Algerian Chronicles, a collection of articles and essays Camus wrote from 1939-1958. Kaplan’s edited edition is the first time these writings have appeared in English, so she is no stranger to Camus and his place in twentieth century French culture.

    Early on Kaplan claims that Looking for the Stranger is actually a biography of Camus’ best known work, and one of the most famous and widely read texts of the twentieth-century. However, this does not mean that Kaplan foregoes a glimpse into Camus’ life, thus resurrecting the Barthesian “death of the author” debate. Instead, Kaplan goes looking for The Stranger in the author instead of the author in The Stranger; the difference is subtly stunning. In other words, her investigation is more preoccupied with the creative process and its cultural and social context than it is with getting to the author as a god-like figure. Camus always claimed that The Stranger was the second in a three part series exploring the absurd from three different perspectives: a novel (The Stranger), a dramatization (Caligula), and a philosophical work (The Myth of Sisyphus). But The Stranger is hardly a book that needs rescuing from obscurity, nor does Kaplan claim that it does. To date the novel has sold over ten million copies and is still read in over 40 languages. It is still on high school and college syllabi, thus making it required reading for young men and women. In fact, a student’s first encounter with existentialism and the absurd is likely to come from a reading of The Stranger. Instead, she offers us a more comprehensive look into the text, running down every lead, exploring every avenue that might expand our understanding of what makes The Stranger the text that it is.

    Kaplan begins by acknowledging the spectacular success of The Stranger, making it one of the most popular and important texts of the twentieth century. She quickly glosses over the critical reaction to The Stranger by pointing out that readings of the novel map some of the most important theoretical lenses that have influenced twentieth century thought. “In fact, you can construct a pretty accurate history of twentieth-century literary criticism by following the successive waves of analysis of The Stranger: existentialism, new criticism, deconstruction, feminism, postcolonial studies” (2). The Stanger, she claims, has influenced thinking of a diverse population that spans generations. Indeed, the remarkable staying power of the novel to remain relevant, perhaps even more relevant now than when it was published, is a feat that its author and its critics at the time could not have foreseen. I am not sure that students continue to read The Stranger with the commitment that they once did, but it is undeniable that the novel still matters, that it still provokes us into thinking, especially in a time when fundamentalism and terrorism are on the rise, and Europe and the United States are flirting with a new form of fascism in the guise of a renewed interest in ridged nationalism. But Kaplan is not necessarily interested in the public and academic reception of The Stranger. Instead, she claims that the novel’s readers and commentators have overlooked something from our reading of the text since its publication: that something is a biography of the novel. “Yet something essential is lacking in our understanding of the author and the book. By concentrating on themes and theories—esthetic, moral, political—critics have taken the very existence of The Stranger for granted” (2-3). She takes the unprecedented, and academically unpopular path that looks into the life of the author and the circumstances that allowed the author at a particular place and time to write one of the most powerful works of world literature. However, it is important to point out that Kaplan sets out to write a biography of the novel, and not the author. In fact, Camus’ life becomes a part of the puzzle that is The Stranger.

    Kaplan is not the first to comment on the unlikely success of The Stranger and its problematic birth. She is, however, the first to devote an entire book to an investigation, an investigation that is almost documentary-like in its approach, to the novel from conception to publication and beyond. And she accomplishes this brilliantly. Told in twenty-six short chapters, bookended by a prologue and an epilogue, Kaplan leads us into the depths of the novel in a highly engaging and thought-provoking fashion. In fact, the structure of her book presents its readers with the “life” of the novel, a life that has continued on long after the death of its creator. Drawing from a reservoir of sources, including Camus’ notebooks and her own trips to Algeria, Looking for the Stranger is a scholarly adventure story. As Kaplan claims in her acknowledgements: “I looked for The Stranger in libraries, in archives, in neighborhoods on three continents” (219). Of course, the idea of The Stranger was with her all of the time, but what makes Kaplan’s book so provocative is precisely the lengths she goes to in search of the novel. Kaplan explores The Stranger in three parts: before its publication, during its publication, and after its publication.

    In the first chapter Kaplan gives us the image of a young man in front of a bonfire burning various papers that link him to a past, a past that could be dangerous to him and those who know him. But as Kaplan tells it, the young Camus could not bring himself to burn all of his letters and writings. What he saved would act as a cache of material, both physical and remembered, from which he would extract and rework into a slim, simply told tale of a man who fails to cry at his mother’s funeral and, by a series of circumstances, ends up shooting an unnamed Arab on a beach, only to be arrested, tried, convicted and sentenced to death. Yet, the reader is never quite sure if the protagonist is convicted and sentenced to death because of the murder or his refusal to conform to the rules of a society that demands that one cry at one’s mother’s funeral. The image of the bonfire given to us by Kaplan is a powerful one. As we travel with her deeper into her investigation, we learn that the bonfire was a kind of rite Camus needed to perform in order to purge his mind and soul so that he could go on to write what he felt needed to be written—unimpeded by ghosts, but still attentive to their silences, which spoke to and through him.

    Throughout the spring of 1940, six years after the bonfire, Camus worked furiously on The Stranger, almost in total isolation holed up in his miserable hotel room in Montmartre, interrupted only to work for five hours a day at Paris-Soir. The twenty-six year old was as cut off from the world as he had ever been. Alone in a foreign city, with German bombs exploding all over France, Camus fought his loneliness and misery by throwing himself into his writing. Not yet divorced from his first wife, Simone Hié, his fiancé Francine Faure refused to accompany him to Paris. The only thing he brought with him was the first chapter of The Stranger and a few of his press clippings. Kaplan: “His sense of separation from everyone he loved put him in a state of mind that was both painful and enabling” (71). Like Camus’ biographer Olivier Todd, Kaplan highlights the importance of Camus’ isolation when he first arrives in Paris. Camus believed that the failure of A Happy Death, his abandoned first novel, was due to his inability to write without interruption. Camus’ isolation in Paris enabled him, out of necessity, to devote all of his attention to The Stranger. Kaplan’s research offers us a marvelous glimpse into the creative process Camus used, or perhaps more accurately, was host to, during his writing of the novel. Kaplan claims that Camus wrote The Stranger almost line for line, as if he were dictating a story he was seeing play out before his eyes. Where he struggled with the writing of A Happy Death, The Stranger seems to have emerged almost fully formed, complete.

    However, his writing of The Stranger does not mean that it was without its problems. In fact, the birth of The Stranger was long and fraught with difficulties both internal and external. Until his arrival in Paris, Camus struggled with getting into the narrative, creating a new story, as well as using material from A Happy Death. Interestingly, most reviewers of Kaplan’s book, Robert Zaretsky, himself an accomplished Camus scholar, and John Williams in particular[v], have devoted a majority of their reviews to the shortage of paper in France as the novel was set to go to press. “To say that the very existence of The Stranger was threatened by the material conditions of the war is no exaggeration, since paper supplies were becoming more and more precious. It looked at one point as if Camus would have to supply his own paper stock!” (136). Camus was in Oran with his family at the time, and was happy to help Gallimard with locating paper. The novel came very close to not being published, but paper stock was found at the last minute and Camus was not obliged to supply his own.

    Once the novel was published it was met with immediate success. But perhaps its success was not so unusual after all. From the beginning Camus wanted the French publishing world, located in Paris, to represent him. In the chapter “A Jealous Teacher and a Generous Comrade,” Kaplan tells the story of Camus’ almost frantic correspondence with Jean Grenier and Pascal Pia, the teacher and the comrade, respectfully, and their influence on The Stranger in its early stages. More importantly, if Camus were to move from a provincial author to a wider audience, one that would include the whole of Europe and possibly America, he would have to seek publication outside of Algeria. As Kaplan notes: “Yet Paris was still the center of book publishing in France, and if Camus wanted to publish outside Algeria, he’d eventually have to find a way to get his manuscript to the capital” (107). This, it seems to me, provides the necessary evidence that Camus was thinking bigger than his native land. He desired a world stage, a stage that would allow his work to be read by the widest possible public and Gallimard was the publisher that could provide him with that opportunity. In his book The Existentialist Moment: The Rise of Sartre as a Public Intellectual, Patrick Baert illustrates the importance of publishing, especially those publishing houses in Paris, for providing the necessary outlet for ideas. “Intellectual ideas spread mainly through publications. Whether through books, magazines, or articles, publishing is central to the rise of intellectual movements. For such movements to be successful, authors have to be well connected to the main publishers and need to have sufficient freedom and power to be able to write what they want to write.”[vi] The network Gallimard could provide Camus with would plug him into some of the most resonant writers and thinkers of the time. As mentioned above, The Stranger was not just a novel, but also an important piece of a longer meditation on the absurd. Therefore, Camus’ relationship with Gallimard, as Kaplan points out, is a key component to his rise to international prominence. Quite frankly, without Gallimard, The Stranger might not have met with its tremendous success.

    Camus’ association with Gallimard was not the only key to his success, however. Gallimard’s star and existentialism’s major voice, Jean-Paul Sartre, also had a lot to do with the success of The Stranger. In his celebrated review of The Stranger, originally published in 1943, Sartre almost single handedly anoints Camus into the French intellectual network, thus solidifying his reputation as a resonant French intellectual. Still, early on in his review Sartre points out that, like its author, The Stranger is a book from “across the sea,” highlighting Camus’ Algerian heritage. Sartre’s generous and insightful review gives a certain intellectual legitimacy to the novel. Sartre: “The Stranger is a classical work, a work of order, written about the absurd and against the absurd.”[vii] This Apollonian form, in the Nietzschean sense, of the novel further reinforces the boundary lines that mark the absurd context, a context that we might fold into the Dionysian, again in the Nietzschean sense.

    But it would be a mistake to consider The Stranger a French novel; it is, in almost every sense, an Algerian novel, a novel obsessed with the sun and the sea. What is perhaps closer to the novel’s intention is, at least in part, a Mediterranean world in a colonial context. In other words, the pied noirs who enjoy French citizenship and the protection it offers as opposed to Arab subjectivity. Arab subjectivity is one of the chief criticisms postcolonial scholars hurl at The Stranger and its author. Yet, a purely postcolonial reading of The Stranger severely limits our understanding of the novel. As David Carroll points out, “I would even say that to judge and indict Camus [as Edward Said does] for his “colonialist ideology” is not to read him; it is not to treat his literary texts in terms of the specific questions they actually raise, the contradictions they confront, and the uncertainties and dilemmas they express. It is not to read them in terms of their narrative strategies and complexity. It is to bring everything back to the same political point and ignore or underplay everything that might complicate or refute such a judgment.”[viii] The postcolonial lens that has dominated readings of The Stranger has also relegated it and its creator to a graveyard for Eurocentric authors. Kaplan’s attention to detail, however, locates the nameless murdered Arab in The Stranger in a central, one might even say, privileged, position. Almost from the beginning, Kaplan admits to being nearly obsessed with the figure of the nameless Arab. Indeed, the namelessness of this character is one of the pivotal points in her book. As Kaplan discovers, there was a nameless Arab in Camus’ life, one that would lead him straight to the central scene in The Stranger.

    In 2015 Other Press published the English translation of Kamel Daoud’s The Meursault Investigation, a retelling of The Stranger from the point of view of the brother of the Arab killed on the beach by Meursault. Daoud, an Algerian journalist living in Oran, writes for Quotidien d’Oran, a French language newspaper in Algeria. The Meursault Investigation is an interesting book that reads more in the style of Camus’ The Fall than The Stranger. The protagonist, speaking to us in the first person from a bar in Oran, informs us that there are other facts in the case that we did not hear, the chief among these is the name of his brother, Meursault’s victim, Musa: “Who was Musa? He was my brother. That’s what I’m getting at. I want to tell you the story Musa was never able to tell. When you opened the door of this bar, you opened a grave, my young friend” (4). Daoud’s text comes dangerously close to being fan fiction. However, there is something profoundly relevant in the novel. The Meursault Investigation demonstrates a deeper understanding of The Stranger, and Camus’ style. In order to write this book, Daoud proves that he knows The Stranger intimately, and his contribution to the story is, indeed, worthy of consideration. The Meursault Investigation demands to be read, digested, and then read again in the context to the cultural as well as the literary conditions of Algeria before, during, and after its independence.

    Kaplan devotes nearly an entire chapter (chapter 26) to Daoud’s novel and the figure of the unnamed Arab who appears in nearly spectral form in The Stranger. She tells us that she had a meeting with Daoud in 2014 in Oran, in which he claimed “we don’t read The Stranger the same way as Americans, French, Algerians” (210).  Kaplan’s reading of Daoud’s novel is a revelatory experience for her, and by association, for us. She strategically situates The Meursault Investigation both within and beyond the lens of postcolonial theory.

    Kaplan’s research into the source of the killing of the Arab scene in The Stranger is a remarkable piece of journalism. Her investigation led her through the towns and alleyways of Oran, to dusty archives, and populated streets, all despite an Algerian travel advisory for those holding a United States passport. “For two years, I had traveled to places in France and Algeria connected to The Stranger: I had walked down the former rue de Lyon in Algiers, past Camus’s childhood home. With photographer Kays Djilali, I climbed the steep Chemin Sidi Brahim, knocking on doors until we found the House Above the World, now the home of three generations of Kabyle women who speak neither French nor Arabic. With Father Guillaume Michel from Glycines Study Center in Algiers, I drove out to gold and blue vistas of Tipasa. In Paris, I stood in the dreary spot on the hill of Montmartre where Camus wrote in solitude” (211). At the end of the trail is a name: Kaddour Touil, and a story.

    Kaplan’s research demonstrates that it is not really Camus the author who haunts The Stranger, but rather it is the specter of Meursault who haunts Camus, both in life and after death. Meursault, as Olivier Todd informs us, is a combination of several people Camus knew. “The character of Meursault was inspired by Camus, Pascal Pia, Pierre Galindo, the Bensoussan brothers, Sauveur Galliero, and Yvonne herself. Marie was not Francine. Camus the writer mastered his novel in a way that Camus the man did not control in his life. Meursault never asked himself any questions, whereas Camus was always examining his actions and motivations.”[ix] Authors routinely use what and who they know for characters and their actions in books, but Camus’ relationship with Meursault seems to be as complicated as that character’s relationship with the reader. Kaplan’s book sheds a new light on the complexities of those relationships.

    The Stranger is truly a work of world literature, in the sense that David Damrosch defines the concept.[x] With The Stranger we have an Algerian author who wrote in French but was influenced by Danish, Russian, and German thinking, and was stylistically influenced by American authors like Hemingway and James M. Cain. Alice Kaplan gives us a view of The Stranger that joins a growing chorus of scholarship on the controversial book and its author. She provides keen insight that opens up other avenues of thinking about that book and its author. Camus’ influence seems to be growing, not diminishing as we move deeper into the twenty-first century, and this is needed, especially given the growing resurgence of nationalism and isolationist polices, i.e. Brexit and Trump. Perhaps it’s only literature, and international fiction in particular, that can save us from ourselves. In this age of social media epitomized by the egotistical selfie, international fiction has become more important than ever. Kaplan’s book reminds us that nothing exists in a vacuum, that great works of art come about contextually and pan-culturally. The Stranger might never have been a success without the French existentialist network of the time.

    Andrew Martino is Professor of English at Southern New Hampshire University where he also directs the University Honors Program. He has published on contemporary literature and is currently finishing a manuscript on the concept of security in the work of Paul Bowles.

    Notes

    [i] Edward W. Said. Culture and Imperialism. (New York: Vintage Books, 1994), 174.

    [ii] Arendt, Hannah. “French Existentialism.” Essays in Understanding: 1930-1954. (New York: Schocken Books, 1994), 192.

    [iii]Albert Camus. The Myth of Sisyphus. Trans. Justin O’Brien. ) New York: Vintage Books, 1991), 30.

    [iv] Randall Collins. The Sociology of Philosophies: A Global Theory of Intellectual Change. (Cambridge, Massachusetts: The Belknap Press of Harvard University Press, 2002), 764.

    [v] See Zaretsky’s review in Los Angeles Review of Books (https://lareviewofbooks.org/article/biography-zaretsky-kaplan-camus/) and Williams’ review in the New York Times (Sept. 15, 2016).

    [vi] Patrick Baert. The Existentialist Moment: The Rise of Sartre as a Public Intellectual. (Cambridge, England: Polity Press, 2015), 138-139.

    [vii] Jean-Paul Sartre. “The Stranger Explained.” We Have Only This Life to Live: The Selected Essays of Jean-Paul Sartre 1939-1975. Ed. Ronald Aronson and Adrian Van Den Hoven. (New York: New York Review Books, 2013), 43.

    [viii] David Carroll. Albert Camus the Algerian: Colonialism, Terrorism, Justice. (New York: Columbia University Press, 2007), 15.

    [ix] Todd, Olivier. Albert Camus: A Life. (New York: Alfred A. Knopf. 1997), 107.

    [x] Here I am thinking specifically of Damrosch’s theory of circulation. See David Damrosch’s What is World Literature. (New Jersey: Princeton University Press, 2003) for a full definition of the concept.

  • Daniel Greene – Digital Dark Matters

    Daniel Greene – Digital Dark Matters

    a review of Simone Browne, Dark Matters: On the Surveillance of Blackness (Duke University Press, 2015)

    by Daniel Greene

    ~

    The Book of Negroes was the first census of black residents of North America. In it, the British military took down the names of some three thousand ex-slaves between April and November of 1783, alongside details of appearance and personality, destination and, if applicable, previous owner. The self-emancipated—some free, some indentured to English or German soldiers—were seeking passage to Canada or Europe, and lobbied the defeated British Loyalists fleeing New York City for their place in the Book. The Book of Negroes thus functioned as “the first government-issued document for state-regulated migration between the United States and Canada that explicitly linked corporeal markers to the right to travel” (67). An index of slave society in turmoil, its data fields were populated with careful gradations of labor power, denoting the value of black life within slave capitalism: “nearly worn out,” “healthy negress,” “stout labourer.”  Much of the data in The Book of Negroes was absorbed from so-called Birch Certificates, issued by a British Brigadier General of that name, which acted as passports certifying the freedom of ex-slaves and their right to travel abroad. The Certificates became evidence submitted by ex-slaves arguing for their inclusion in the Book of Negroes, and became sites of contention for those slave-owners looking to reclaim people they saw as property.

    If, as Simone Browne argues in Dark Matters: On the Surveillance of Blackness, “the Book of Negroes [was] a searchable database for the future tracking of those listed in it” (83), the details of preparing, editing, monitoring, sorting and circulating these data become direct matters of (black) life and death. Ex-slaves would fight for their legibility within the system through their use of Birch Certificates and the like; but they had often arrived in New York in the first place through a series of fights to remain illegible to the “many start-ups in slave-catching” that arose to do the work of distant slavers. Aliases, costumes, forged documents and the like were on the one hand used to remain invisible to the surveillance mechanisms geared towards capture, and on the other hand used to become visible to the surveillance mechanisms—like the Book—that could potentially offer freedom. Those ex-slaves who failed to appear as the right sort of data were effectively “put on a no-sail list” (68), and either held in New York City or re-rendered into property and delivered back to the slave-owner.

    Start-ups, passports, no-sail lists, databases: These may appear anachronistic at first, modern technological thinking out of sync with colonial America. But Browne deploys these labels with care and precision, like much else in this remarkable book. Dark Matters reframes our contemporary thinking about surveillance, and digital media more broadly, through a simple question with challenging answers: What if our mental map of the global surveillance apparatus began not with 9/11 but with the slave ship? Surveillance is considered here not as a specific technological development but a practice of tracking people and putting them into place. Browne demonstrates how certain people have long been imagined as out of place and that technologies of control and order were developed in order to diagnose, map, and correct these conditions: “Surveillance is nothing new to black folks. It is a fact of antiblackness” (10). That this ”fact” is often invisible even in our studies of surveillance and digital media more broadly speaks, perversely, to the power of white supremacy to structure our vision of the world. Browne’s apparent anachronisms make stranger the techniques of surveillance with which we are familiar, revealing the dark matter that has structured their use and development this whole time. Difficult to visualize, Browne shows us how to trace this dark matter through its effects: the ordering of people into place, and the escape from that order through “freedom acts” of obfuscation, sabotage, and trickery.

    This then is a book about new (and very old) methods of research in surveillance studies in particular, and digital studies in general, centered in black studies—particularly the work of critical theorists of race such as Saidiya Hartman and Sylvia Wynter who find in chattel slavery a prototypical modernity. More broadly, it is a book about new ways of engaging with our technocultural present, centered in the black diasporic experience of slavery and its afterlife. Frantz Fanon is a key figure throughout. Browne introduces us to her own approach through an early reflection on the revolutionary philosopher’s dying days in Washington, DC, overcome with paranoia over the very real surveillance to which he suspected he was subjected. Browne’s FOIA requests to the CIA regarding their tracking of Fanon during his time at the National Institutes of Health Clinical Center returned only a newspaper clipping, a book review, and a heavily redacted FBI memo reporting on Fanon’s travels. So she digs further into the archive, finding in Fanon’s lectures at the University of Tunis, delivered in the late 1950s after being expelled from Algeria by French colonial authorities, a critical exploration of policing and surveillance. Fanon’s psychiatric imagination, granting such visceral connection between white supremacist institutions and lived black experience in The Wretched of the Earth, here addresses the new techniques of ‘control by quantification’—punch clocks, time sheets, phone taps, and CCTV—in factories and department stores, and the alienation engendered in the surveilled.

    Browne’s recovery of this work grounds a creative extension of Fanon’s thinking into surveillance practices and surveillance studies. From his concept of “epidermalization”—“the imposition of race on the body” (7)—Browne builds a theory of racializing surveillance. Like many other key terms in Dark Matters, this names an empirical phenomenon—the crafting of racial boundaries through tracking and monitoring—and critiques the “absented presence” (13) of race in surveillance studies. Its opposition is found in dark sousveillance, a revision of Steve Mann’s term for watching the watchers that, again, describes both the freedom acts of black folks against a visual field saturated with racism, as well as an epistemology capable of perceiving, studying, and deconstructing apparatuses of racial surveillance.

    Each chapter of Dark Matters presents a different archive of racializing surveillance paired with reflections on black cultural production Browne reads as dark sousveillance. At each turn, Browne encourages us to see in slavery and its afterlife new modes of control, old ways of studying them, and potential paths of resistance. Her most direct critique of surveillance studies comes in Chapter 1’s precise exegesis of the key ideas that emerge from reading Jeremy Bentham’s plans for the Panopticon and Foucault’s study of it—the signal archive and theory of the field—against the plans for the slave ship Brookes. It turns out Bentham travelled on a ship transporting slaves during the trip where he sketched out the Panopticon, a model penitentiary wherein, through the clever use of lights, mirrors, and partitions, prisoners are totally isolated from one another and never sure whether they are being monitored or not. The archetype for modern power as self-discipline is thus nurtured, counter to its own telling, alongside sovereign violence. Browne’s reading of archives from the slave ship, the auction block, and the plantation reveal the careful biopolitics that created “blackness as a saleable commodity in the Western Hemisphere” (42). She asks how “the view from ‘under the hatches’” of Bentham’s Turkish ship, transporting, in his words, “18 young negresses (slaves),” might change our narrative about the emergence of disciplinary power and the modern management of life as a resource. It becomes clear that the power to instill self-governance through surveillance did not subordinate but rather partnered with the brutal spectacle of sovereign power that was intended to educate enslaved people on the limits of their humanity. This correction to the Foucauldian narrative is sorely necessary in a field, and a general political conversation about surveillance, that too often focuses on the technical novelty of drones, to give one example, without a connection to a generation learning to fear the skies.

    Stowage of the British slave ship Brookes under the regulated slave trade act of 1788
    “Stowage of the British slave ship Brookes under the regulated slave trade act of 1788.” Illustration. 1788. Library of Congress Rare Book and Special Collections Division Washington, D.C.

    These sorts of theoretical course corrections are among the most valuable lessons in Dark Matters. There is fastidious empirical work here, particularly in Chapter 2’s exploration of the Book of Negroes and colonial New York’s lantern laws requiring all black and indigenous people to bear lights after dark. But this empirical work is not the book’s focus, nor its main promise. That promise comes in prompting new empirical and political questions about how we see surveillance and what it means, and for whom, through an archaeology of black life under surveillance (indeed, Chapter 4, on airport surveillance, is the one I find weakest largely because it abandons this archaeological technique and focuses wholly on the present). Chapter 1’s reading of Charles William Tait’s prescriptions for slave management, for example, is part of a broader turn in the study of the history of capitalism where the roots of modern business practices like data-driven human resource management are traced to the supposedly pre-modern slave economy. Chapter 3’s assertion that slave branding “was a biometric technology…a measure of slavery’s making, marking, and marketing of the black subject as commodity” (91) does similar work, making strange the contemporary security technologies that purport the reveal racial truths which unwilling subjects do not give up. Facial recognition technologies and other biometrics are calibrated based on what Browne calls a “prototypical whiteness…privileged in enrollment, measurement, and recognition processes…reliant upon dark matter for its own meaning” (162). Particularly in the context of border control, these default settings reveal the calculations built into our security technologies regarding who “counts” enough to be recognized. Calculations grounded in an unceasing desire for new means with which to draw clear-cut racial boundaries.

    The point here is not that a direct line of technological development can be drawn from brands to facial recognition or from lanterns to ankle bracelets. Rather, if racism, as Ruth Wilson Gilmore argues, is “the state-sanctioned or extralegal production and exploitation of group-differentiated vulnerability to premature death,” then what Browne points to are methods of group differentiation, the means by which the value of black lives are calculated and how those calculations are stored, transmitted, and concretized in institutional life. If Browne’s cultural studies approach neglects a sustained empirical engagement with a particular mode of racializing surveillance—say, the uneven geography produced by the Fugitive Slave Act, mentioned in passing in relation to “start-ups in slave catching”—it is because she has taken on the unenviable task of shifting the focus of whole fields to dark matter previously ignored, opening a series of doors through which readers can glimpse the technologies that make race.

    Here then is a space cleared for surveillance studies, and digital studies more broadly, in an historical moment when so many are loudly proclaiming that Black Lives Matter, when the dark sousveillance of smartphone recordings has made the violence of institutional racism impossible to ignore. Work in digital studies has readily and repeatedly unearthed the capitalist imperatives built into our phones, feeds, and friends lists. Shoshanna Zuboff’s recent work on “surveillance capitalism” is perhaps a bellwether here: a rich theorization of the data accumulation imperative that transforms intra-capitalist competition, the nature of the contract, and the paths of everyday life. But her account of the growth of an extractive data economy that leads to a Big Other of behavior modification does not so far have a place for race.

    This is not a call on my part to sprinkle a missing ingredient atop a shoddy analysis in order to check a box. Zuboff is critiqued here precisely because she is among our most thoughtful, careful critics of contemporary capitalism. Rather, Browne’s account of surveillance capitalism—though she does not call it that—shows that race does not need to be introduced to the critical frame from outside. That dark matter has always been present, shaping what is visible even if it goes unseen itself. This manifests in at least two ways in Zuboff’s critique of the Big Other. First, her critique of Google’s accumulation of  “data exhaust” is framed primarily as a ‘pull’ of ever more sites and sensors into Google’s maw, passively given up users. But there is a great deal of “push” here as well. The accumulation of consumable data also occurs through the very human work of solving CAPTCHAs and scanning books. The latter is the subject of an iconic photo that shows the brown hand of a Google Books scanner—a low-wage subcontractor, index finger wrapped in plastic to avoid cuts from a day of page-turning—caught on a scanned page. Second, for Zuboff part of the frightening novelty of Google’s data extraction regime is its “formal indifference” to individual users, as well as to existing legal regimes that might impede the extraction of population-scale data. This, she argues, stands in marked contrast to the midcentury capitalist regimes which embraced a degree of democracy in order to prop up both political legitimacy and effective demand. But this was a democratic compromise limited in time and space. Extractive capitalist regimes of the past and present, including those producing the conflict minerals so necessary for hardware running Google services, have been marked by, at best, formal indifference in the North to conditions in the South. An analysis of surveillance capitalism’s struggle for hegemony would be greatly enriched by a consideration of how industrial capitalism legitimated itself in the metropole at the expense of the colony. Nor is this racial-economic dynamic and its political legitimation purely a cross-continental concern. US prisons have long extracted value from the incarcerated, racialized as second-class citizens. Today this practice continues, but surveillance technologies like ankle bracelets extend this extraction beyond prison walls, often at parolees’ expense.

    A Google Books scanner’s hand
    A Google Books scanner’s hand, caught working on WEB Du Bois’ The Souls of Black Folk. Via The Art of Google Books.

    Capitalism has always, as Browne’s notes on plantation surveillance make clear, been racial capitalism. Capital enters the world arrayed in the blood of primitive accumulation, and reproduces itself in part through the violent differentiation of labor powers. While the accumulation imperative has long been accepted as a value shaping media’s design and use, it is unfortunate that race has largely entered the frame of digital studies, and particularly, as Jessie Daniels argues, internet studies, through a study of either racial variables (e.g., “race” inheres to the body of the nonwhite person and causes other social phenomena) or racial identities (e.g., race is represented through minority cultural production, racism is produced through individual prejudice). There are perhaps good institutional reasons for this framing, owing to disciplinary training and the like, beyond the colorblind political ethic of much contemporary liberalism. But it has left us without digital stories of race (although there are certainly exceptions, particularly in the work of writers like Lisa Nakamura and her collaborators), perceived to be a niche concern, on par with our digital stories of capitalism—much less digital stories of racial capitalism.

    Browne provides a path forward for a study of race and technology more attuned to institutions and structures, to the long shadows old violence casts on our daily, digital lives. This slim, rich book is ultimately a reflection on method, on learning new ways to see. “Technology is made of people!” is where so many of our critiques end, discovering, once again, the values we build into machines. This is where Dark Matters begins. And it proceeds through slave ships, databases, branding irons, iris scanners, airports, and fingerprints to map the built project of racism and the work it takes to pass unnoticed in those halls or steal the map and draw something else entirely.

    _____

    Daniel Greene holds a PhD in American Studies from the University of Maryland. He is currently a Postdoctoral Researcher with the Social Media Collective at Microsoft Research, studying the future of work and the future of unemployment. He lives online at dmgreene.net.

    Back to the essay