The Anxious Age – Is Post-Postmodernity Defined by Mental Distress?

2013-07-02 17.34.55 - Copy

 

 

Introduction: Countermodernism

 

In 1979 Jean-François Lyotard (1924-1998) over-simplistically defined Postmodernism as “incredulity toward metanarratives, which was undoubtedly a product of progress in the sciences”.[1]  More simplistically still, this means that the theoretical machines of Modernism and before were rapidly ceasing to function as science progressed.  In Lyotard’s words, “postmodern scientific knowledge cannot be based on performativity, because efficiency must be calculated based on a stable system”. Nature and society are not stable (or closed) systems, and as such it is counter-intuitive to evaluate and accept universal truths derived from the variables of those systems. The truth-value in a stable system is inherently homogeneous and tyrannical – it will always favour the knowledge-bearer and work towards hegemonic disparity (knowledge is power).  Metanarratives, then, are told from a position of privilege or State, and any events of knowledge-based ontogenesis have their relative successes and progressions judged against a minor system of inconsistencies and innovations.  “New moves,” as Lyotard called them. As is demonstrated by dictatorships, control does not promote innovation, but rather homogenises the system.  This echoes in many ways Gilles Deleuze (1925-1995) and Felix Guattaris’ (1930-1992) analogy of Chess and Go whereby the game of chess represents the state, while Go epitomises the War Machine (the difference between a closed or stable system and an open and unstable system).[2]  Knowledge in the postmodern world is about change, adapting to it, and generating new ideas (nomadology), not on an established rigid scientific method. In 2019 we are indeed sceptical of grand narratives, yet our faith in science is perhaps greater than ever.  Mental illnesses, which are too often ascribed to biological causes, are medicated with confidence by general practitioners, and ingested blindly by the patient.

In Lyotard’s own terms, this precarious era following Postmodernism may then be defined as a blind credulousness toward the metanarrative of the sciences.  This will neither be proven nor disproven as, like Lyotard, I invite and anticipate paralogy (Lyotard again, referring to a discourse without consensus – an open system), though I do hope to add my voice to the already rich – somewhat cacophonous – well-pool of perspectives to be found in contemporary theory.

No discourse is ever initiated truly objectively: the writer always has a vested interest in the points made and the results found, contrary to any claim against this.  For is it not the initial spark of enthusiasm for any topic which drives that writer onwards, and is this not in and of itself self-interest?  My own vested interests are first-person perceptions of parenting a child whose behaviour is erratic, violent, recalcitrant and without context.  I propose to show how this behaviour, which is increasingly commonplace, carries not just biological hallmarks but is also linguistic in nature.  This perspective is parallel to Mark Fisher’s (1968-2017) belief that mental illness (in particular Bipolar Disorder) are the consequence of life in a neoliberal world rather than a biological imbalance, and as such it is vital to the sprit of this study to compare Postmodern theorists such as Deleuze and Guattari (who maintained that culture is biological) with the latter-day writings of Fisher.

This study takes the singular hybrid form of a third-person autobiography which will jump in and out of first-person perspective and critical analysis.  It does this because in order to evaluate the minor narrative it behoves us, in the manner of Postmodernism, to also evaluate the metanarrative.  In short, while I will be questioning the sources of my information, it will be equally important to evaluate my own narrative – to question the contingencies which have occasioned my experiences and shaped my perspectives thereon.

img006 - Copy

This is by no means the first attempt I have made to link our present social, political and cultural epoch to mental health.  In 2016 I wrote:

 

“(t)he schizophrenic is subject to fragmented thinking and delusions, synthesises words which make only subjective sense to the patient; repeats words and phrases over and over, each time as if for the first.  The schizophrenic displays a lack of emotional expressions, shows little to no enthusiasm and exhibits repetitive, jarring speech abnormalities.  Studies have shown that a key environmental factor in the onset of schizophrenia is childhood separation or loss – dislocation from a previous generation.  The post-millennial condition is schizophrenic in all of these factors and more.  Perhaps the most critical similarity, though, is in the delusional impersonation of established personalities of import without prejudice to the historical or mythological frameworks in which they belong.”[3]

 

If the term “schizophrenic” is appropriate to our times – if indeed the metanarrative is, in itself, schizophrenic – then it is the state which promotes a fractured society.  Not only, then is this a cultural matter – it is universal in its scope and implications.  Let us begin with the following statement, and allow this study to take us where it may:

 

“10% of children and young people (aged 5-16 years) have a clinically diagnosable mental problem, yet 70% of children and adolescents who experience mental health problems have not had appropriate interventions at a sufficiently early age.”[4]

 

What does this extract from a 2008 study suggest about our present age?  First, it speaks of a huge disparity between mental illness and our understanding of it.   Secondly, and more perniciously, it can only mean that healthcare professionals can no longer keep up with the vast amount of caseloads which grow exponentially by the year.  Numbers alone cannot provide or subtract weight to a hypothesis – if there are 10,000 children, and 10 percent of those have mental illnesses, out of those 1,000 children 300 have had sufficient intervention, while 700 have not.  And how do we qualify the sufficiency of “sufficient intervention?”  Has it been sufficient to bring the young person back from the brink of crisis and prevent self-harm or suicide, or has it been sufficient to provide them with the intellectual and rational tools to live lives relatively free of such crises?  Furthermore, the word “diagnosable” invites all manner of discourse, for is not science here limited to a pre-established metanarrative?  Psychiatry and paediatrics follow strict diagnostic criteria built on decades-old research, focused all-but-exclusively on biological study.  The linguistic question in such studies is a secondary concern, while any study of psychological abnormalities in our present age must, by necessity, bring linguistic and environmental factors more into focus.

It is for this very reason that I eschew statistical data: it belongs to the fixed sciences, those closed systems which are regulated by serotonin and its parameters of activity.

 

V__2B85

 

 

 

Episode I: PostNormal Hyper-Reality

 

The trouble begins in 1992 at the age of 14.  Nothing out of the ordinary happens that year – no bereavement, no stress…no perceivable external reason for the boy to make an attempt on his own life.  The only warning sign is a gradual onset of ennui, a sense of hopelessness and despair the like of which the boy’s parents are at a loss to account for.  One evening it occurs to him that the most sensible, rational act would be to overdose on pain killers, go for a walk and allow the drugs to destroy his organs.  When he wakes up in the hospital, the boy is overcome with the certainty that this would be the first of many such occurrences.  Over the years the boy will have no less than six hospital admissions on his medical record with varying degrees of seriousness.

 

In hindsight the boy (quite patently a reactivated prosopopoeia of my younger self) recalls that 1992 was likely one of the first periods of his life when the full pernicious implications of the neoliberal orthodoxy.  That year we marched in London in protest of the implementation of Poll Tax, and that same year (and contrary to previous legislation to abolish it), it was announced that Poll Tax would be replaced by Council Tax in 1993 (Thus the age of rebranding was born which continues to the present day).

 

The 1990s saw the dominance of the SSRI (Selective Serotonin Reuptake Inhibitor) over the field of mental health and psychiatric treatment.  Commonly referred to in the US by the brand name Zoloft and in the UK as Lustral, Sertraline is the most commonly prescribed SSRI, replacing Prozac in the early Twenty-First Century as the by-word for antidepressants.  I first began taking Sertraline in 2002, but it proved no difficulty in finding someone who had a more recent experience with the drug.  This user, who wished to remain anonymous, likened the first sensation as euphoric: “slightly drunk, without the loss of inhibition.”   The subject also went on to suggest that there may be placebic elements to SSRIs: “after the initial rush of happiness, it felt like I was carrying on the course of medication just to avoid withdrawal – which the doctor warned me would be unpleasant.”

 

Perhaps we are looking at antidepressant medication in the wrong way.  Instead of providing a means for coping with the world around us, could it not be the case that SSRIs actually create docile bodies?  Slavoj Žižek famously claimed about John Carpenter’s They Live,

 

“…definitely one of the forgotten masterpieces of the Hollywood Left. … The sunglasses function like a critique of ideology. They allow you to see the real message beneath all the propaganda, glitz, posters and so on. … When you put the sunglasses on you see the dictatorship in democracy, the invisible order which sustains your apparent freedom.”[5]

14141785_10210525902487691_4241852284410737561_n

Sertraline, and medication which acts in similar ways, act as the sunglasses in this example, yet the effect is opposite.  To put the sunglasses on is to become blind to the insidious processes of control and power which govern society.  A cursory glance at the listed side-effects of Sertraline seem to confirm this:

 

  • depression, feeling strange, nightmare, anxiety, agitation, nervousness, decreased

sexual interest, teeth grinding,

  • shaking, muscular movement problems (such as moving a lot, tense muscles, difficulty

walking and stiffness, spasms and involuntary movements of muscles), numbness and

tingling, abnormal taste, lack of attention,

  • visual disturbance, ringing in ears,
  • palpitations, hot flush, yawning,

 

As a long-term user of Sertraline, I can attest to all of these side-effects, and more.  The decreased sexual interest is of particular importance, having a Lacanian frame of reference.  Is this not the virtual definition of castration?  To repress desire in this way is to dissolve the Oedipal Stage, revealing the “real” father – which in this case is the State itself, and we can place more stock in Lacan here than perhaps any other contemporary theorist. If we contrast the figure of authority between that of a century (or even half-a-century) ago to the figure of authority of today, the difference is vast.  What was once domestic and proximal (the literal father) is now global and distal (the state apparatus), and this latter has no need to even interact directly with us: it has proven much more effective for us to actively regulate ourselves.  We surrender our desires to the desires of ideology, allowing the neoliberal clinamen to prevail.  While it is no great revelation to say that governments and ruling elites control society in ever-more pernicious ways (proving this would indeed be akin to proving that water is wet), there is also the neoliberal concept of “post-truth” to contend with which paints our political climate in colours far brighter than those of reality.  Anthropologist Alexei Yurchak coined the term “hypernormalisation” to describe the attitude of paradoxical political blindness which permeated Russia in the latter years of Soviet rule[6], and the progression from this to what Mark Fisher called “Capitalist Realism” is self-evident: while the former pretends that the climate is functional, the latter knows that the opposite is true, yet cannot imagine an alternative.  Fisher in fact takes his cue from the State Realism of the Soviet Union and its propagandist machinery, Realism has nothing to do with the Real.  On the contrary, the Real is what realism has continually to suppress.”  SSRIs are the perfect societal adjunct to this state – what better way of maintaining the illusion of stability than freely giving subjects the perception-managing drugs they crave?

 

221808_2012316711440_6520795_n

 

 

Episode II: Psychomodernism Vs. Schizomodernism

 

The child screams.  Perfunctory sounds no longer suffice to convey the ever-more complex thought processes in his head, nor do the articulated phonics which he has learned to parrot back at his father allow these abstractions to manifest themselves.  In his frustration the child begins to slam his head against the living room wall, once, twice…until the very succession of this action has deadened the images in his mind that he cannot yet begin to grasp.  The skin on his forehead is aflame with pain and the wall retains the crimson memory of the boy’s rage.  The child’s terrified father sobs as he applies the towel-wrapped frozen bag of vegetables grabbed in a panic from the fridge to his son’s head.  As these incidents increase in both frequency and severity, the child’s parents naturally seek medical help, only to find themselves subjected to the most tortuous cross-examinations and intense scrutiny.

 

This is, without question, a linguistic problem.  And, sure enough, as the child’s vocabulary becomes more sophisticated, so the violent head-banging decreases.  This would, ordinarily, serve as the happy ending scenario to a troubling-yet-not-altogether-atypical parental crisis.  However, over the course of the following years more troubling symptoms develop: an explosive aversion to fire alarms, the sound of a hand dryer in a public toilet, an intolerance for clothing…all of these point towards behaviour typical of the autism spectrum.  At 7 years old the child is removed from his school following an attempt to strangle a classmate, and what follows is a nearly two-year diagnostic period during which the child is assessed by numerous professionals in order to gauge speech and language (to satisfy the linguistic question), psychology (to address mental health) and paediatrics (for autism-related issues).

 

There are also the protracted and focused attacks by the child upon his mother, resulting in the latter’s body (in particular her lower torso) being covered in bruises and swelling.  Occasionally these attacks are facial, and black eyes become common.  And then there are the secondary effects.  The child’s mother is forced to forego employment during the week, surrender her studies (ironically in social care) and the relationship between her and the child’s father eventually becomes so jaded and warped that they end up finding one another again at the end of the process – only this time they are entirely different individuals to the couple who fell in love a decade previously: they are beaten, disillusioned…all promise of future prosperity scuppered by bureaucratic torpor.

 

It is now seven years since the child began to display troubling behaviour, and two years since his parents sought help from the Umbrella Pathway, a service provided by Worcestershire County Council to “provide an assessment process for all children and young people presenting with neuro-developmental disorders which may be due to Autism Spectrum conditions (not ADHD).”[7]  Among the non-diagnostic suggestions made by professionals is PDA:

 

“Pathological demand avoidance (PDA) is a behavioural profile associated with apparently obsessive non-compliance, distress, and florid challenging and socially inappropriate behaviour in children, adolescents and adults.”[8]

 

While PDA is a behaviour profile within the autism spectrum, it is by no means unanimously agreed upon by professionals whether or not it belongs on the autism spectrum.  It is therefore referred to as a sub-type.

 

When I informed my older brother of this prospective diagnosis, his immediate response was to exclaim “you’ve just described yourself!”  Could it then be that there is a biological element which I have passed onto my son, which has only become recognised scientifically in his generation?  In hindsight I recall my childhood carrying hallmarks of PDA: an aversion to authority, discomfort at regulation and intense feelings of suppressed rage.182634_1880487455791_3649772_n

However, what if the problem lies elsewhere, in the most pernicious and overlooked social evil: standardisation?  A hallmark of neoliberalism, standardisation regulates the mainstream of the state apparatus, covering all areas of government and the public/private sector.

 

The government’s own website states:

 

 

Standardisation is the process of creating, issuing and implementing standards. A standard is a document, established by consensus and approved by a recognised body. It provides rules, guidelines or characteristics for activities or their results so that they can be repeated. They aim to achieve the greatest degree of order in a given context.[9]

 

The two words which jump out there are “order” and “repeated.”  These hallmarks of meta-power can be traced to Deleuze and Guattaris’ notion of the state apparatus and the war machine and to Michel Foucault, who would say that “order” in this context can be translated into “discipline” in order to produce normalisation and therefore “docile subjects.”

 

Episode III: Dromomodernism and Aggressive Desublimation (There Can Be No Conclusion)

 

In Precarious Rhapsody, Franco Berardi states:

 

“The acceleration of information exchange has produced and is producing an effect of a pathological type on the individual human mind and even more on the collective mind. Individuals are not in a position to process the immense and always growing mass of information that enters their computers, their cell phones, their television screens, their electronic diaries and their heads. However, it seems indispensable to follow, recognise, evaluate, process all this information if you want to be efficient, competitive, victorious. … The necessary time for paying attention to the fluxes of information is lacking.”[10]

 

In a hypernormalised world of post-truth, what better way to control a people than to bombard them with a constant strobe of information parcels?  As Twitter feeds accelerate and Facebook becomes ever-more hyperbolic, so too do our levels of anxiety.  Can PDA be rooted in biological neurosis which is exacerbated by linguistic factors?  Paul Virilio (1932-2018) argued that “there was no ‘industrial revolution’, only ‘dromocratic revolution’; there is no democracy, only dromocracy; there is no strategy, only dromology.”[11]  Dromology is derived from the Greek “dromo,” which refers to the activity of racing, ergo speed and acceleration.  Dromology, then, is surely how we should countenance the flow of information in the modern age. There can be no true conclusion to this study: as linguistic cultural and social factors multiply and accelerate, we can only wait to see how our biological and linguistic bodies cope…if, indeed, they can.

 

16864141_10212279478926006_5779708440142606051_n

 

 

[1] Lyotard, J., Bennington, G., Massumi, B. and Jameson, F. (2005). The postmodern condition. Manchester: Manchester University Press.

[2] Deleuze, G., Guattari, F. and Massumi, B. (2017). A thousand plateaus. London: Bloomsbury Academic. (pages 523-551)

[3] Davis, G. (2016). No Job for a Grown Man (part six) – in Explication of the Schizophrenic Age. [online] Legally, I Own the Thoughts of the Dead. Available at: https://grumpusart.wordpress.com/2016/11/07/no-job-for-a-grown-man-part-six-in-explication-of-the-schizophrenic-age/ [Accessed 30 Apr. 2019].

[4] Children’s Society (2008) The Good Childhood Inquiry: health research evidence. London: Children’s Society.

[5] Slavoj Žižek, THE PERVERT’S GUIDE TO IDEOLOGY. British Board of Film Classification (19 June 2013).

[6] Jurchak, A. (2006). Everything was forever, until it was no more. Princeton, NJ: Princeton University Press.

[7] Hacw.nhs.uk. (2019). Umbrella Pathway. [online] Available at: https://www.hacw.nhs.uk/our-services/childrens-community-health-services/umbrella-pathway [Accessed 30 Apr. 2019].

[8] (Newson et al. 2003; O’Nions et al. 2014b)

[9] GOV.UK. (2019). Standardisation. [online] Available at: https://www.gov.uk/government/publications/standardisation/standardisation [Accessed 30 Apr. 2019].

[10] Berardi, F. (2010). Precarious rhapsody. London: Minor Compositions.

[11] Virilio, P. and Polizzotti, M. (2006). Speed and politics. New York: Semiotext(e).

Advertisements

Through the (Immediate) Past, Darkly

We are compelled to bookend the event, to portion months and years up into manageable, quantifiable volumes of matter and memory.  If time is the ultimate capitalist commodity, then our quantification of time is its currency, and it is in this way that the annual review-of-the-year rundowns which one can read in any given broadsheet or tabloid, or viewed on December 31st through a gaze of varying levels of ridicule are -to all intents and purposes – its audit.  Is it of any value to do this from a philosophical angle?  And, indeed, wherein lies the point?  Is it in order to file away each successive year into a unitary index for the historian or sociologist to access at their convenience?  If so, are we not further commodifying our notions of time?  In order to give this interrogation some perspective, perhaps we should benefit from referring to Manuel DeLanda (1952 – )’s excellent introduction to A Thousand Years of Non-Linear History:

 

“(I)f the different “stages” of human history were indeed brought about by phase transitions, then they are not “stages” at all – that is, progressive developmental steps, each better than the previous one, and indeed leaving the previous one behind.  On the contrary, much as water’s solid, liquid, and gas phases may coexist, so each new human phase simply added itself to the other ones, coexisting and interacting with them without leaving them in the past.”[1]

time-and-free-will-an-essay-on-the-immediate-data-of-consciousness

This being considered, is not our Gregorian inclination towards sectioning off units of duration rendered utterly meaningless?  Certainly, it may serve to “time-map” individual and collective events, it can also be useful as measurement of progress and decline.  It can, however, be of very little use to the contemporary thinker as Élan vital.  It is futile to review a year in terms of its singularity.  What we think of as “time” is little more than the shifting of energy, the passing of matter from one state to another.  There was, and never could be a 2018, as duration, that which we think of as “time” would necessitate each antecedent year forcing themselves as one into that year which is being experienced.  This is Bergsonism at its purest (albeit, too, at its most simplified), and if we were to take that model of the  present being nothing other than the past happening all at once – insofar as one may interpret such a complex idea out of its parole – then we may find a perfect analogy for our times: the Twenty-First Century can be defined by its concentrated repetition of the past, both culturally and politically, and perhaps provides us with the first significant parallels between Henri Bergson (1859 – 1941) – for decades dismissed as an antiquity of the old guard in philosophy – and Marxist poststructural thinkers such as Jacques Derrida (1930 – 2004) and, more recently, Mark Fisher (1968 – 2017).  For, if we were to take the concept of Bergson’s Duration out of the confines of analytic philosophy and place it in the broader spectrum of Critical Theory (and we surely can, as was amply proven by Gilles Deleuze (1925 – 1995)’s re-interpretations of Bergsonism), then what we are presented with is a like-for-like match of what Fisher called The Slow Cancellation of the Future[2], the “temporal malaise” which is so much a hallmark of contemporary culture that one is hard-pressed to discern between that which has been created last week and the artefacts of the 1970s and 1980s.

slow cancellation 1

 

Contemporary thinking, I would suggest, has all but eradicated the notion of a sole Philosopher King stood atop his plateau (to borrow the Deleuzian analogy).  This has been the case for several decades.  In his 1957 foreword to the second edition of Critique of Everyday Life, Henri Lefebvre (1901 – 1991) notes

 

“…professional philosophers generally ignored the book; for – starting with its title – it entailed relinquishing the traditional image of the philosopher as master and ruler of existence, witness and judge of life from the outside, enthroned above the masses, above the moments lost in triviality, ‘distinguished’ by an attitude and a distance.” [3]

 

Philosophy has, over the years, necessarily been a process of cross-pollination of thought.  In Elemental Discourses, John Sallis (1938 – ) writes

 

(i)n Derrida’s texts there are many voices. Some occur as citations from Husserl, Heidegger, or other authors. Yet, in the strict sense whatever is set forth in citations is not the voice of another but rather a passage from a written text. Even if what is cited should happen to be words once heard in the voice of another, they will, in being cited, have been transposed into the written text; in this transposition the voice of the other will have been silenced. And yet, we sometimes attest that in reading the words of an author we can hear his voice behind the words, that we can hear it silently resounding.”[4]

 

Thus we may observe the clinamen of Lucretius evolve throughout the ages and become Deleuze and Guattaris’ desire, accounting for the clinamen’s inclination towards capital.  Indeed, the most pertinent and evolutionary use of philosophy is to commandeer from its massive historical inventory of themes and ideas – and, it can be argued, this is how philosophy finds its true meaning (Deleuze and Guattari themselves said “the only question is which other machine the literary machine can be plugged into, must be plugged into in order to work.  Kleist and a mad war machine, Kafka and a most extraordinary bureaucratic machine ….”[5]).  One need not absorb every text by Foucault, or attend two-hour lectures on Lacan to gain a healthy reserve of critical resources with which to formulate one’s own theories. Plato’s cave and Wittgenstein’s stonemasons serve very well as building blocks for an architecture of language and socialisation, reality and simulacra.  These ideas the modern philosopher must osmose and re-interpret, modify and apply pressure to, and for that very reason philosophical models function in much the same way as art: intense critical thought and complex abstractions simplified to the nth degree as signs, giving flesh to otherwise untranslatable concepts: art builds real architecture in Utopia and peoples it accordingly, yet it draws its strength from its ability to topple said architecture and rebuild.  Artistic movements provides the zeitgeist for this architecture, and these zeitgeists are the very agents of its destruction and reformation.  Much to Plato’s imagined chagrin, art is in many ways inseparable from critical thinking. At any given moment, the human mind is subject to incalculable heterogonous abstractions which superficially bear no relation to one another other than their chronological linearity – or the oft-cited stream of consciousness, that convenient one-size-fits-all coat with which lazy commentators have dressed such diverse literary figures as Beckett, Burroughs, Thompson, Joyce and Proust. Terms such as stream of consciousness exist to categorise that which has no formal category (other than, in this instance, that of literature).  But, if we again consult Bergson, thought processes are time in its purest state.

 

“Let us assume that all the sheep in the flock are identical; they differ at least by the position which they occupy in space, otherwise they would not form a flock. But now let us even set aside the fifty sheep themselves and retain only the idea of them. Either we include them all in the same image, and it follows as a necessary consequence that we place them side by side in an ideal space, or else we repeat fifty times in succession the image of a single one, and in that case it does seem, indeed, that the series lies in duration rather than in space. But we shall soon find out that it cannot be so. For if we picture to ourselves each of the sheep in the flock in succession and separately, we shall never have to do with more than a single sheep. In order that the number should go on increasing in proportion as we advance, we must retain the successive images and set them alongside each of the new units which we picture to ourselves: now, it is in space that such a juxtaposition takes place and not in pure duration.”[6]

 

The Twenty-First Century has, since 2001, been bereft of landmark political or social moments.  The key word here is “landmark,” indicating a fixed point in time after which the ideological apparatus in place before the event can no longer function, such is the impact it has on society, economics and culture.  The word “landmark” also implies space, rather than time, yet is no misuse: chronology and geography are intermingled in memory, creating those very ghosts which populate Derrida’s hauntology, and in keeping with the concept of hauntology, the most critical phenomena of the year occurred just as it was ending.  Two separate and distinct events, which happened no more than a week from one another at the end of December and superficially bear little-to-no relation, but which in fact have great reciprocal significance.  Firstly, Charlie Brooker’s Black Mirror series gave us another instalment in the form of the feature-length Bandersnatch, which was closely followed by the announcement that HMV had gone into second liquidation, His Master’s Voice now nothing but a pitiable whimper in the neoliberal wind.  As outmoded a capitalist model as it is out of touch with the times, HMV has, for decades, pre-packaged culture and sold it on as part of some great promise that what that culture represents is the very essence of what one needs to understand our times.

28-black-mirror-bandersnatch-2.w700.h467

 

Within the first few minutes of Bandersnatch it becomes apparent that popular culture will never tire of revisiting the 1980s, as though that decade was both the genesis and zenith of our postmodern metanarrative.  And yet again, the past is shown to us through countless factual and technical filters – for instance, it is safe to say that nobody ever bought a Tangerine Dream album in WHSmith in the mid-1980s.  WHSmith, like HMV represents the Harrods model of “everything under one roof,” which for a store that deals in entertainment and culture, is a laughably hyperbolic claim.  Yet our memories of these shops, for those of us who had childhoods in the 1980s, portray them as precisely that, for our own undeveloped awareness of the sheer richness and variety of culture is reflected by HMV’s own limited scope of same.  Thus, it adequately met our stunted expectations.  As culture and technology evolved at an ever-increasing rate in the latter half of the Twentieth Century, it soon became the case that this capitalist model of the third place serving as cultural nexus could never fulfil our ever-more-sophisticated understandings of culture.  This could partially explain our craving for nostalgia, as the artefacts of the 1980s remind us of the last time we were culturally satiated – the economy of craving and fulfilment was in balance (perhaps it is only in childhood that this balance is ever truly equal).  “Nostalgia,” though, as Simon Reynolds (1963 – ) points out, is translated etymologically as “homesickness.”[7]   Contemporary sociologists favour the notion of a fourth place in order to tackle the workplace/home environment crossover, but it is more accurate to re-identify The Third Place as increasingly virtual.  This is hardly surprising, since 9/11 shattered what was quite possibly the West’s final moment when an event was experienced collectively in The Third Place, and was the last “where were you when…?” moment in living memory.  As the Twenty-First Century has unfolded, global events have occurred in what feels like a steady trickle, owing not to a lack of event, but in the way in which events are now relayed to us.  In the sixteen years since the towers collapsed, the ingestion of current affairs has gradually slipped away from the static television screen and become something experienced singularly (one-on-one) through portable, streamlined devices.  Before the internet, the news was fed to us daily at precise quarters of a clock, with the 6 and 9 PM instalments reserved for in-depth investigations into the ramifications of the day’s events.  This may well still be the case, but it is now by no means how we initially learn of these events, which are continuously fed to us via the offices of internet newsfeeds which have no beginning or end, and wholesale information dumps such as Twitter.  News is no longer dropped on us four times a day around a centralised information hub (i.e. television or radio), but is now with us all day, and can be accessed from any location via mobile phones, tablet and laptops.  Wi-Fi has freed us from the necessity of the specific location, and thus the “where were you?” moment can no longer really exist, since such an occasion is marked by more quotidian, tangential social interactions (since social media, we are paradoxically no longer social beings) – history has always been made in conjunction with analogue discourse to provide context and understanding; the pause for reflection has been superseded by the knee-jerk re-tweet.  Is it any wonder, then, that cultural eruptions comparable to that of 1976 have been scarce-to-non-existent during the last decade-and-a-half, and that the cultural satellites of the punk movement can now be bought in Primark on t-shirt racks which also contain images of Miles Davis and Marvel superheroes?

In this sense we can quite easily relate the lack of modern social information exchanges and their replacement by personalised feeds of information to a Twenty-First Century flatness, or to put it simply, an age when globally-relevant events are still unfolding on a daily basis but are no longer felt as shockwaves.  Without shockwaves there can be no fissures, which is where Twentieth-Century culture once thrived: jazz, pop, punk, Abstract Expressionism, Conceptual Art and the Postmodern break in general all happened as a consequence of events which were felt as they occurred, and carried real consequences, unlike the political pantomimes of today.  The ages in which these events happened had their own zeitgeist modelled from the social mood, and are remembered – perhaps rightly or wrongly – for their cultural and social values.  In an age which has had no real shockwaves or fissures a void has inevitably been created which has no atmosphere, zeitgeist or – crucially – human analogue.  Since domestic concerns are primarily centred around economy, the average Western citizen concerns him or herself with financial survival and the waning scope for prosperity.  Let us, for a moment, contemplate upon a strata of people working not for prosperity or an elevated standard of living, but only in order to cling onto the standard of living they already have.  The opiate of the masses has been superseded by a cold bucket of water, terror at a knock upon the door.  The clinamen of capital (its desire) is absolute subjectification.  In 2018, Brexit once more proved itself to be that very subjectification, spreading fear and hatred across the UK and using similar (albeit more sophisticated) tools to divide the country as did Germany in 1939.  Brexit is more pernicious than the campaigns of history, however: there is no single identified common enemy, no one sub-section of society singled out for persecution.  Rather, it plays to the worst fears of all social stratifications, always with one lingering threat – you will lose what you have. Again, as with the trial of Adolf Eichmann in 1961, one can discern a palpable sense of what Hannah Arendt called “the banality of evil,” a clear attitude of “rather you than me.”  Brexit has played into the neoliberal ideology in the only way it could: divide and conquer.  But in recent years, it has become the norm for events to resemble past situations.  Occupy, it can be argued, was itself a modified sit-in, grafted from the late-1960s onto the present day and given an Economics degree.  Where it has prospered – as opposed to the disenfranchised, disconnected youth of fifty years ago – is in its organisation and the clarity of its voice, both of which can be attributed to technological agencies unimaginable in the last century.  We have lost our sense of the epoch-making event, the galvanising force to attempt something different: the rule book is no longer torn up, so much as it is re-told through post-millennial perspectives.

wwi-color-restored

October saw the release of Peter Jackson’s They shall Not Grow Old, a technically astonishing colourised documentary to mark the centenary of Armistice Day.  Nothing can detract from the visual and journalistic achievement, although one could also read the film via Jean Baudrillard and liken the process to, for instance, the endeavours of Japan’s Ōtsuka Museum of Art, where only precise facsimiles of well-known original works are displayed.  The museum is, of course, anything but that: indeed, one might say that it is part-PowerPoint presentation / part-PT Barnum grotesquerie.  Or perhaps the film is more akin to the Abbey of Saint-Michel-de-Cuxa which was rebuilt by John D. Rockefeller, Jr.in Upper Manhattan using building materials from the original abbey in Southern France from what remained after it was rebuilt in 840.  The latter suggestion is given more weight when one considers that much of the original footage found in They Shall Not Grow Old was filmed using arcane hand-crank cameras which struggled to maintain a steady 12 frames-per-second, which is why the original films appear so jerky.  This jerkiness has been digitally offset by high-end digital trickery to save the World War One soldiers from an eternity of coming across like “…Charlie Chaplin-type figures.[8]”  This obviously means that fifty percent of what one sees in Jackson’s film is not original footage at all, but very sophisticated computer animation.

When Jackson says “I wanted to reach through the fog of time and pull these men into the modern world, so they can regain their humanity once more” he misses the mark somewhat for what has actually happened is that these veterans have indeed been brought out of the past, but only in a digital suspension: half-human, half-computer-generated chimera, they hang on the screen like exhibits from the Ōtsuka Museum.  Similarly, much of the audio track to They Shall Not Grow Old is actors’ dialogue, translated via the offices of a deft lip-reader who no doubt spent as many hundreds of hours reviewing the original footage as Jackson’s team did animating it.

Curiously, one can with great ease finish watching They Shall Not Grow Old and immediately begin watching the first episode of Peaky Blinders (set in 1919, one year after armistice) without any disruption in either narrative or visual quality.  Peaky Blinders itself, is an example of history’s reworking and re-presentation.  Its non-diegetic soundtrack is entirely of the Twenty-First Century, and consists of artists aping the late 1970s and early 1980s.

 

There can be no table of contents for 2018, nor can it be reviewed month-by-month.  The writer cannot simply disclose a year as a series of events which range in importance or ramification.  I certainly have not done this (nor would I ever wish to).  Charlie Brooker, when not writing episodes of Black Mirror, will scan the year in a linear manner in his New Year’s Eve Wipe, but for serious discourse this can never fully articulate the essence of a twelve-month duration.  I will, however, borrow one recurring sound-off from Brooker:

 

“That was (2018)…now go away.”

 

[1] DeLanda, M. (1997). A Thousand Years of Nonlinear History. New York: Zone Books, pp.15-16.

[2] Fisher, M. (2014). Ghosts of my life. Winchester: Zero Books, pp.21-39.

[3] Lefebvre, H. (1991). Critique of everyday life. 2nd ed. London: Verso, p.5.

[4] Sallis, J. (2018). Elemental Discourses. Bloomington: Indiana University Press, p.13.

[5] Deleuze, G. and Guattari, F. (2004). A thousand plateaus. London: Continuum, p.5.

[6] Bergson, H., Ansell-Pearson, K. and Ó Maoilearca, J. (2002). Key writings. New York: Continuum, pp.49-50.

[7] Reynolds, S. (2012). Retromania. London: Faber and Faber, p.50.

[8] Ilse, J. (2019). Prince William attends World Premiere of “They Shall Not Grow Old”. [online] Royal Central. Available at: http://royalcentral.co.uk/uk/cambridges/prince-william-attends-world-premiere-of-they-shall-not-grow-old-110509 [Accessed 2 Jan. 2019].

Two Steps Back: a Critique of Today, a Dismissal of the Past and a Eulogy for the Future, as Presided over by Mark E. Smith and The Fall

“I had come to loathe my husband, Mr Harlax.  I mean, physically, be revolted by him.  I could look at him and think only of the functions.”

 

  • Artemis ‘81

 

Différance is the funeral held for the meta-narrative.

 

Memorex.  Manufacturer of computer peripherals and recordable media.  Established in 1961, Memorex were synonymous during the 1980s with home recording.

 

Kraken.  Legendary squid-like sea creature said to wrap its tentacles, once disturbed, around vessels and drag them to the bottom of the ocean.

 

The time of year I remember most distinctly from my childhood were those strange weeks when the nights drew in.  Halloween, Bonfire Night…the cheap masks at the shop at the end of the twitchel (because that’s what they were called in North Nottinghamshire), the divine aroma of potatoes being charred on the backyard fire which, in our age of ultra-safety, would never be bureaucratically tolerated.  Those cold, dark evenings carried their own gothic magic as a child.  One could quite easily imagine Spring-Heeled Jack bounding from the council estate roofs and the bizarrely-gnarled trees in the woods actually being science fiction organisms.  Renowned as one of the most haunted villages in England, the remains of an old Roman garrison sat atop the clay hill which hamletted the village on all sides.  There was always a spectral threat on the lips of our parents, and all of this has indelibly left a quasi-Victorian gothic impression on my recollections of the early eighties.

 

To begin, then, with the problematic word.  When we say “haunting,” we are tacitly referring to the ineffable: concepts which, when attempted to give form to or study, vaporise.  Something altogether apart from philosophical immanence, this is the run-out groove which carries the fading analogue vibrations of our specific pasts, and if words such as “haunting,” “ghosts,” “spectres,” (ad nauseum) are to characterise memory this is only because these terms serve best to outline a difference which cannot be described in binaries.  We may, if we are so inclined, steer off track and cite Bergson at this point though it serves just as well to propose that memories are recalled in units, rather than successive elements of time.  Were we to recall perfectly our entire lives in reverse beginning with the absolute present then we would doubtless pick up on subtle ideological or cosmetic shifts in our environs.  We would, however in all probability miss the greater shifts and distinctions, but given that this kind of recollection is impossible, we instead focus on the event.  These events, as unitary measures, are themselves “haunted,” as it were, by dead elements (be they cars which are no longer on the road, a foodstuff no longer manufactured or a television programme that nobody else remembers being aired).  We may then say that we are haunted by the event, or even the unit.

 

This is hinted at by Derrida, yet made explicit by the 21st Century permutation of Hauntology.  Our factually oblique and rose-tinted recollections of the past coupled with the present conundrum of  “already been done” has suspended Western culture in a temporal loop.  “Two Steps Back,” in fact.  The time-locked cultural blockage of an age after Postmodernism has rendered the “new” profoundly spectral: we are watching, listening and responding to ghosts.  These ghosts are the spectres of Modernism and pre-Modernism, the last cultural epochs where technological and biological growth were anywhere near in tandem with one another.  Indeed, the “new” is necessarily enshrouded with quotation marks – even visually, the word is spectral.  Unknown to me in the 1980s, the literal ghosts alluded to in local folklore were in fact the unconscious parental responses to a time which made more harmonious sense, when there were less technological leaps to bemoan.

 

Différance is the individuation between biological memory, political memory, cultural memory…it is how Proust’s memories distinguish themselves between Dostoyevsky’s, how Beckett’s memories are internalised whereas Joyce’s remain geographical.  Escape from Marienbad, indeed.  Différance is the temporal linguistic rift rent by dromology. Différance is the colloquial vapour trails left hanging in the air in the wake of cultural imperialism.

 

Différance is the family unit with lost unity.

 

Derrida likened the spectre of Marx (that phantasmagorical after-image which has haunted capitalism for over a century) to Hamlet’s father: literally and etymologically the root spectre at the feast.  Indeed, that crucial textual link was made early on by highlighting that Hamlet was “the Prince of a rotten state,” allegorically that same rotten state which was to be found in the wake of communism itself, its ectoplasmic remains congealed in the scattered debris of the Berlin Wall.  1980s Britain, or its working class communities, had more than it’s share of this rot: alcoholism, redundancy, solvent abuse, domestic violence, mass unemployability…all of these were to be found, as a child, beneath the superficial halcyon sheen of the nuclear family.

 

 

 

IMG_5678

This impression is what always returns when one hears The Fall.  The oblique, rumbling production on Dragnet, the keyboard trail on Frightened, the choppy vaudeville of City Hobgoblins.  And those words…like tapping into long-forgotten truths which revealed themselves in layers the more one could discern them.  Listening to any Fall record was worth a dozen trips to the library and provided a far more comprehensive (albeit labyrinthine) education than one could hope to gain in those Thatcherite penal colonies we were forced to attend during the week: instant psychic Cinerama of a world made up of grotesque (ha!) dog-breeders, phantom stalkers, Disneyland beheadings and strange conjugations of literary figures.  Mark E. Smith saw himself as a writer above all else, and it is indeed within those wordscapes that one is ensnared once those primitive, repetitive rhythms and snarling Northern barks have either enchanted or repelled you.  One reads The Fall as one reads Deleuze – in layers and multiplicity; the libido in despair, castrated by its own production.  Listen to Room to Live, or Tempo House and you have a Deleuzian machine absorbing as it creates.  One can almost hear the ideas forming in Smith’s mind just before he contorts them, the rhythm section in endless repetition as time strangles the pleasure principle.

 

Once one hears The Fall (either as a joyous or attritional experience) one is at once haunted by The Fall: like Marxism, the time between first contact and present time is rotten with phantoms.  The “ghosting” effect on an old television broadcast is merely the ghost of multiplicity, information forced down a tube which is continuously caught up with itself in a cathode Möbius.  The “captured” cultural elements of the past, ensnared by Smith, become distorted in much the same way as Francis Bacon would pervert his subjects and, like Bacon, Smith froze his subjects at their most primal as though intuition led him to their animal state: Terry Waite, Alan Minter, MR James, Lou Reed and Doug Yule (in an instant fused into the one chimeric state) – all in a state of “…becoming Fall.”  The industrial landscapes sonically conjured by a superficially grotesque rumble are another “becoming,” for in that instantly primal cacophony lay not only the bleak Conservative wasteland of late-70s and early 80s, but also admitted to the industry of Blake’s Jerusalem – a bleakness far sootier and rooted in diaspora than anything suggested by Kraftwerk or Joy Division.  Here was (and is, captured in essentia) an industry transcending political trend: if Marxism is the spectre haunting Europe (macro), then The Fall conjure the specificity of a Britain enslaved to a Marxist ontology, or rather the phantom of differance which manifested itself in a typically Northern blue-collar attitude which eternally defies translation.

IMG_5758.JPG

This was the Britain one would experience if one watched Coronation Street on through a lens in any way similar to Smith’s – the Barlows’ crepuscular killing sprees, Kevin Webster copulating with Jack Duckworth’s pigeons in the outhouse to produce a malformed beak/moustache hybrid, all in those lurid cathode reds and blues of early colour television, yet with shadows darker than a Castiglione monoprint.  And we respond to those grotesqueries knowing full well that we – the working class with our fathers risking life and limb daily at the colliery – are the grotesque products of a perverted society.  Smith took the narrative experimentation of The Velvet Underground and twisted it to his own vision, throwing in all manner of literary, cultural and political allusion along with it – the mystical autodidact Roman Totale XVIII his early prosopopoeial alter-ego emerging from the song lyrics to commandeer the sleeve notes.  So within, so without.

 

Music has been something which comes and goes in my life, with precious few exceptions.  When I wanted to put together a band at the age of twelve, I was too young to do so to any extent other than drafting in school friends to help create an undisciplined cacophony in the spare room.  It was, for them, something to stave off boredom and nothing more.  For me, it quickly became the case that I could entertain myself more effectively by making a cheap guitar and amplifier sound like something other than a cheap guitar and amplifier: hiding the amp under a pillow with the bass turned all the way up sounded like the atonal hum of a building site, whilst striking the strings with metallic objects made for the sounds of cinematic stabbing (reflecting the potentially lethal act of striking electric guitar strings with metallic objects in the first place).  I made no further attempt to form a band until in my early twenties, when any chance of finding like-minded individuals had been scuppered by the lumpen musical ideals left in the wake of a withered Brit Pop: young musicians wanted to sound like solo Paul Wellers and already-existing bands stank of being fully endorsed by their parents, who were probably in bands which emulated Paul Weller in 1983.  There was nowhere to be found the kind of person who wanted to create the kind of sonic experimentation I needed to make.  The Velvet Underground’s currency – which has always been in fluctuation in the eyes of the mainstream – was at an ebb.  Such is the way with being out-of-synch with things: always one paradigm shift from having a real chance at something special.  In hindsight, of course, I count my blessings for the music business is among the vilest of industries, and I then lacked the sheer bloody-mindedness to persist at all costs like Mark E. Smith always has.  More than this, I lacked the discipline to maintain a single sound throughout anything approaching a career, much less stick to so singular a discipline as music.  This is why the likes of Smith, Billy Childish and other counter-culture luminaries who doggedly refuse to attenuate themselves to anything as crass as a marketable sound have to formulate their own economies – and not just financial economies.

 

Shane Meadows’ 1999 drama A Room for Romeo Brass was filmed in the same village alluded to in the first paragraph.  Shot nine years after I left the village, there is a marked difference in the landscape of my childhood and that recreated on the screen, a difference which went beyond representation.  Seeing the village in Meadows’ film, I felt no nostalgia, no sudden desire to return there.  Indeed, aside from the novelty of recognition, there was nothing to link the me in the present to the me who recalled playing in the exact locations now being used as a stage for Paddy Considine.  Partially, this can be attributed to simple displacement and the passage of time, but more crucially the topography had altered to such an extent over those nine short years that my very conception of the village had become the recollection of a ghost, or at the very least an erasure.  My childhood existed only in my memory, and no amount of old photographs (of which there are very few) could ever amount to anything more than a multiplicity of reflection.  Time is no longer a thing which can be measured by temporality alone – of all the images and zeitgeists left to us by the Twentieth Century, a sense of echoing pastiche is likely the dominant sensation which has only increased with massive exponentiality to the present day.  Which decade is this year in the 1990s emulating?  To what extent do the purveyors of culture in 2012 understand the forms and aesthetics they are aping from 1969?

 

I have, since an age too far back in my memory to place with any exactitude, been in a state of mourning.  This is no silly Freudian claim of being desirous of a return to the womb: personally, I frequently refer to that oft-repeated Smiths lyric whenever I encounter Freud – “it says nothing to me about my life.”  The mourning I claim is the mourning for a childhood half received, or indeed a deferral of childhood which was felt just as (if not more than) keenly during my infancy.  Betraying the above claim, I must nonetheless turn to Freud for his unheimlich to describe that jarring notion as a child that there was always something wrong, something awry or missing.  Unheimlich is perversely the most fitting term for my domestic childhood situation, for the home was sporadically and decidedly unhomely.  Growing up with alcoholism from an extremely early age means that there has been no chance for the child to know anything other than a home run through with alcoholism, and that home being in a relatively (by today’s standards) tight-knit community means that any social comparisons must be drawn from other homes which are in some way complicit with alcoholism (few could have not known that our house was the one with the parent who lapsed wildly into stupors lasting days and, sometimes, weeks.  Yet very little was ever done to circumvent the vicious circle of dependency: in fact, the reverse was so often the case).  In such circumstances, one lives in a microcosm of Other: there is nothing wrong with this picture…and everyone who knows precisely what is not wrong with the picture knows how to mind their own business about what is not wrong – at least until their front door is closed.

 

When an infant encounters an adult who is drunk, the first instinct is to think of the adult as “unwell,” which is conveniently confirmed by other adults and becomes the official euphemism. “Unwell” also means “absent” in such cases, even if the unwell person is in the same room, because the sober parent has been purged of all parental virtues, such as responsibility, kindness, indulgence or accommodation.  The entire architecture of home life is dismantled to such an extent that the very state of childhood is placed in suspension.  If a parent is too drunk to collect their child from nursery, then that child ceases being a child in the eyes of nursery staff and becomes a problem.  If a child is not in school because of parental alcoholism, then that child is now a “case.”

 

But what is perhaps the most destructive of all are those periods when the parent is sober: life is less complicated, certainly, and the parent/child bond is soldered together once more, but there is always the dread – which can occur at any time, with or without warning or cause – that the unwell will return, rendering the moments of sobriety something to fear just as much as the periods of chaos.  This, then, is the mourning I have felt since infancy.  Petite morts in the most literal sense: mourning the death of the home, the death of a childhood being allowed to live itself out, the small, staggered death of a parent.

 

I was six years old in 1984, the year forever burned into scholarly discourse as the official death of the blue-collar worker in Great Britain.  My father worked in the mines, though ours was a colliery who outlasted many others in Nottinghamshire.  Although Calverton Colliery almost survived the century due to private finance (the office block was the last structure to be demolished in March, 2000), redundancy hit our household in 1987-1988.  The Calverton in A Room for Romeo Brass is the neoliberal perversion of industry: Vicky McClure’s character works in a fashion outlet in St. Wilfred’s Square, which was formerly a chemist; Considine’s Morrell is unemployed, friendless and entirely disconnected from both morality and self, a parody of identity tripping over itself to fill in the cracks left over from a patchy education and a (tacitly) fractured home life.  This is a society very much in the process of restructuring itself, redefining its identity by drawing from its immediate past, its discordant present and its bleak future.  Almost twenty years on and that future is not so much bleak as eerie: children no longer play in the streets (an ostensibly glib statement at first glance, but no less true for it).  For children in 2018, socialising has become compartmentalised into school, after-school clubs and birthday parties.  The common ground of having parents who worked within a location-specific industry is gone, and in its place are streets full of adults who are too busy keeping their heads above water with insufficient McWages to integrate with others on the street – often because those others command higher wages for less effort, but perhaps more often because there is little understanding of what the other’s job actually entails.  What once unified communities has alienated it.  Mark Fisher adroitly pointed to this diasporic labour culture as both cause and symptom of depression in his excellent essay Good for Nothing.  The working-class curse of being made to feel inadequate for professional jobs, whilst feeling inferior (or at the very least, fraudulent) in office or factory work:

 

“…because I was overeducated and useless, taking the job of someone who needed and deserved it more than I did. Even when I was on a psychiatric ward, I felt I was not really depressed—I was only simulating the condition in order to avoid work, or in the infernally paradoxical logic of depression, I was simulating it in order to conceal the fact that I was not capable of working, and that there was no place at all for me in society.”

 

It is only natural that this phenomenon trickles down to our children.  As displaced as we are in our present economy, this can be nothing compared to a child who feels no tangible connection to a world both virtual and indifferent.  We recognise ourselves (or partial vestiges of ourselves) in our immediate culture and react to this accordingly, yet when our immediate culture is purely virtual (such as is the case when a child’s daily routine consists of school-dinner-bathtime-device, as opposed to a routine from the late Twentieth Century which was more akin to school-play-play-dinner-play-bathtime-bed), psychic well-being suffers just as surely as physical well-being suffers from vitamin deficiency.  Children identify with – and, terrifyingly, become – nebulous, uncanny forms in video games: forms which have nondescript facial characteristics, limited movement and lifespans with no value.  They are both Geppetto and Pinocchio with no reference to a higher meaning.  Small wonder, then, that they struggle to place any real value to the social realm.  The mirror stage ceases to function when the mirror ceases to reflect.  We are now (and have been for some time) in an age of mass childhood dysfunction which has increased at such exponential speed that psychologists, behaviourists, therapists (et al) can no longer sufficiently account for it.  This is because the accounting must come from fields outside of Freudian specificity: the social sciences (as evidenced by Fisher) are where the answers are to be found, and from a sociological perspective they are to be found relatively easily.  We need only refer to Foucault and thereby note the homogenisation of the state apparatus (the school being modelled on the prison, for example) to see the link between this and a gaming platform such as Roblox, which takes this model to the nth degree.  Players create and interact in virtual cell-like buildings, which can vary between prisons, schools, houses, pizzerias or indeed any simulacrum of our reality.  Very little distinguishes these artifices aside from superficial décor, and the tasks each player performs is largely standardised and based on production / consumption.  The neoliberal ideal supplied (as only the neoliberal ideology would be allowed to) as plaything for a standardised socialising.  If any suggestion had been made to me (and, I imagine, any other child) in the 1980s that performing perfunctory tasks in order to achieve virtual (i.e. non-existent) rewards could in any way be passed off as entertainment or – even more scandalously – playtime, this would have been dismissed as some species of Stalinism: a tin-pot attempt at coercing child labour masquerading as fun.  Platforms such as Roblox offer no conduits for the superego to develop, and creativity is limited to the basic additions the child can make to their domains.  This is the very business school one imagines when listening to The Birmingham School of Business School from the 1992 album Code:Selfish:

 

 

Weave a web so magnificent

Disguise in the art of conceit

….

Deposits prisoner robotics

Home to their wives Stepford

Case-carrying

Business School

 

At the expense of all else the neoliberal worldview must emulate itself, asserting its financial and political dominance in self-replication, deceit and a means to an end mentality (the end of which must ever be kept out of sight and grasp).

 

IMG_6086

I was twenty-five when my mother died, just one week after her sixtieth birthday.  Those small, staggered periods of mourning I had undergone all throughout my life until that point returned, massively intensified and furiously indignant at the torment I had lived through.  To have my mother’s death played out in front of me so many countless times, whereby the person who should have been a constant in my life mockingly replaced by something so animalistic finally and so swiftly taken from me at a point in my own life when I should have been adjusting and reacting to the vicissitudes of my own adulthood felt like the most vicious betrayal of all.  Depression had been a factor in my life since the age of twelve (if I have to give an age to the time when it was finally recognised that the sense of “wrongness” at home had finally been absorbed by my own psyche to become an unwellness in its own right), and by the time of my mother’s death I had already made no less than six attempts at my own life.  Any attempts at academia up to that point were offset or sabotaged by personal feelings of insufficiency and I had tellingly fallen into catering – a vocation frequently associated with verbal abuse and physical suffering.  All relationships I had were a priori doomed to failure, though that only served to exacerbate the pain when this inevitably became the case.  Again, the protracted mourning period playing itself out.

 

A Memorex, then, for the Krakens.  These memories remain buried, submerged beneath countless quotidian events waiting to be re-activated by sensory stimuli.  The stimuli, though, must be of the time of the memory in order to function.  The Memorex must be a pure recording.

 

Ti West’s 2007 film The House of the Devil goes further than pastiche: it wants you to believe that it was made in the early 1980s, down to the camera tints, synth-heavy soundtrack, dialogue and content (devil worshippers here deliberately chosen to harken back to the Satanic Panic in the wake of the Richard Ramirez killings).  Most tellingly, however, is the film’s title shot.  Filling half the screen in garish yellow, the title reeks of cheap exploitation horror though the inclusion of the film’s date in Roman numerals gives pause: the tradition of placing the film’s title with it’s production date directly underneath with All Rights Reserved is something which died out in the late 1970s, thus creating not only a jarring anachronism but also – perhaps most poignantly – turning the charade in on itself.  What we are left with is not a reference to the past, but rather an atemporal, half-remembered throwback which forfeits historical exactitude in favour of nostalgia for a time which never happened as it exists in collective memory.  The House of the Devil is by no means alone in this stylised misappropriation: It Follows, Beyond the Black Rainbow, The Neon Demon, Nightcrawler, Under the Skin and Amer are but a small selection from the hundreds of motion pictures made with an eye to providing the viewer with that most ultra-postmodern thrill of experiencing the past as they have always remembered it: not factually, but mnemonically via associations and cultural connection.  The danger of this, of course, is in the potential for collective memory to wipe out the historical fact.  A Twenty-One-year-old watching these films today has no first-hand experience of 1984, therefore leaving them with nothing to distinguish between the two oppositions.  Reason concludes that the result of this phenomena will be an entirely muddled collective memory in 40-50 years whereby the Twentieth Century will eventually be remembered amorphously and atemporally.

ice_screenshot_20180521-042744

Though the above may be a peculiarly altermodernist symptom, its eventual effect is, to all intents and purposes, what one is listening to on And This Day, Hex Enduction Hour’s cataclysmic denouement.  Time crashes in on the listener all at once, the preceding fifty-three minutes of the album serving as individual elements while And This Day serves them all up at once.

In 2018, we are still in the process of mourning (a deferred mourning, but a mourning nonetheless). We mourn the failed promises of modernism while we adjust constantly to the increasing pressures of a neoliberal world.  For the working classes, we mourn ourselves as we struggle to ward off the demands of abstract capital.  Our ongoing mental and psychic collapse is as much the product of Victorian Dad ideology as it is lagging concentration in an age of advanced dromology.  “Pull your socks up” is scandalously still being uttered by mental health workers who themselves cannot ever hope to reach the bottom of the piles of cases stacking up every day.  Those children lucky enough to be dealt with in timely fashion are furnished with ADHD statements as readily as birth certificates, while other children less fortunate (mine included) wait years to be granted a cursory inspection, before an inevitable non-conclusive conclusion.  The fault lies squarely with the parents, so the official party line of responsibilisation goes.  Parents, however, are sinking under ever-increasing debt just to stay above water.  For the working classes, the very concept of a meritocracy is not as ludicrous as it is offensive. Perhaps this penchant (yearning, even) for the relics of the past – albeit reformatted to fit in with our collective memory – is nothing less than a coping strategy: there was a time when those in need would be accommodated, when the poor were dealt with sympathetically rather than with scorn.  And as much as we know this to be far from the truth, it is a falsehood far more comfortable than today’s crushing truths.  Mark E. Smith was the ever-present rage against the horrors of neoliberalism: fiercely opposed to the fol-de-rol of social media and distrusting to the end of a system which streamlines cultural endeavour to fit the device, Smith took The Fall and made it rougher as the rest of the world became sleeker.  The grotesque salmagundi of sound sculpted in the 1980s, consisting as it did of harsh Germanic repetition, quasi-Jamaican barked ad-libbing, Velvet Underground drone and a brash form of working class country music (country and northern, if you will [and he did]) had, over the last decade, become a feral beast of unrelenting curmudgeonly fury, primed and aimed at any and all facet of a West so utterly surrendered to the growing weight of capital.

 

Mon coure et je suis d'accord

As amusing as it may be to recall Smith’s innumerable bon mots, jibes and drunken slurs collected over the decades, it is nonetheless to miss the point – Samuel Beckett was no less the caustic wit when in his frequent cups and Jackson Pollock could just as easily clear a dinner party as Smith could a pub.  Yes, I frequently return to YouTube for my regular fix of Smith’s brusque humour in interviews yet, for the proper stuff, I delve feet-first into Grotesque and Hex Enduction Hour.  These albums weren’t joking.  They meant every rancorous syllable.  While Morrissey was regaling us with upturned bicycles and Oscar Wilde throwbacks, Smith gave us the world red in tooth and claw, only redder and toothier.  And while the former produced countless soundalikes throughout the eighties, nineties and to this day, nobody has ever managed to sound like The Fall.  Quite right, too.