The Clockwork Penguin

Daniel Binns is a media theorist and filmmaker tinkering with the weird edges of technology, storytelling, and screen culture. He is the author of Material Media-Making in the Digital Age and currently writes about posthuman poetics, glitchy machines, and speculative media worlds.

Category: Media

  • Elusive images

    Generated with Leonardo.Ai, prompts by me.

    Up until this year, AI-generated video was something of a white whale for tech developers. Early experiments resulted in janky-looking acid dream GIFs; vaguely recognisable frames and figures, but nothing in terms of consistent, logical motion. Then things started to get a little, or rather a lot, better. Through constant experimentation and development, the nerds (and I use this term in a nice way) managed to get the machines (and I use this term in a knowingly reductive way) to produce little videos that could have been clips from a film or a human-made animation. To reduce thousands of hours of math and programming into a pithy quotable, the key was this: they encoded time.

    RunwayML and Leonardo.Ai are probably the current forerunners in the space, allowing text-to-image-to-(short)video as a seamless user-driven process. RunwayML also offers text-to-audio generation, which you can then use to generate an animated avatar speaking those words; this avatar can be yourself, another real human, a generated image, or something else entirely. There’s also Pika, Genmo and many others offering variations on this theme.

    Earlier this year, OpenAI announced Sora, their video generation tool. One assumes this will be built into ChatGPT, the chatbot which is serving as the interface for other OpenAI products like DALL-E and custom GPTs. The published results of Sora are pretty staggering, though it’s an open secret that these samples were chosen from many not-so-great results. Critics have also noted that even the supposed exemplars have their flaws. Similar things were said about image generators only a few years ago, though, so one assumes that the current state of things is the worst it will ever be.

    Creators are now experimenting with AI films. The aforementioned RunwayML is currently running their second AI Film Festival in New York. Many AI films are little better than abstract pieces that lack the dynamism and consideration to be called even avant-garde. However, there are a handful that manage to transcend their technical origins. But how this is not true of all media, all art, manages to elude critics and commentators, and worst of all, my fellow scholars.

    It is currently possible, of course, to use AI tools to generate most components, and even to compile found footage into a complete video. But this is an unreliable method that offers little of the creative control that filmmakers might wish for. Creators employ an infinite variety of different tools, workflows, and methods. The simplest might prompt ChatGPT with an idea, ask for a fleshed-out treatment, and then use other tools to generate or source audiovisual material that the user then edits in software like Resolve, Final Cut or Premiere. Others build on this post-production workflow by generating music with Suno or Udio; or they might compose music themselves and have it played by an AI band or orchestra.

    As with everything, though, the tools don’t matter. If the finished product doesn’t have a coherent narrative, theme, or idea, it remains a muddle of modes and outputs that offers nothing to the viewer. ChatGPT may generate some poetic ideas on a theme for you, but you still have to do the cognitive work of fleshing that out, sourcing your media, arranging that media (or guiding a tool to do it for you). Depending on what you cede to the machine, you may or may not be happy with the result — cue more refining, revisiting, more processing, more thinking.

    AI can probably replace us humans for low-stakes media-making, sure. Copywriting, social media ads and posts, the nebulous corporate guff that comprises most of the dead internet. For AI video, the missing component of the formula was time. But for AI film, time-based AI media of any meaning or consequence, encoding time was just the beginning.

    AI media won’t last as a genre or format. Call that wild speculation if you like, but I’m pretty confident in stating it. AI media isn’t a fad, though, I think, in the same ways that blockchain and NFTs were. AI media is showing itself to be a capable content creator and creative collaborator; events like the AI Film Festival are how these tools test and prove themselves in this regard. To choose a handy analogue, the original ‘film’ — celluloid exposed to light to capture an image — still exists. But that format is distinct from film as a form. It’s distinct from film as a cultural idea. From film as a meme or filter. Film, somehow, remains a complex cultural assemblage of technical, social, material and cultural phenomena. Following that historical logic, I don’t think AI media will last in its current technical or cultural form. That’s not to say we shouldn’t be on it right now: quite the opposite, in fact. But to do that, don’t look to the past, or to textbooks, or even to people like me, to be honest. Look to the true creators: the tinkerers, the experimenters, what Apple might once have called the crazy ones.

    Creators and artists have always pushed the boundaries, have always guessed at what matters and what doesn’t, have always shared those guesses with the rest of us. Invariably, those guesses miss some of the mark, but taken collectively they give a good sense of a probable direction. That instinct to take wild stabs is something that LLMs, even a General Artificial Intelligence, will never be truly capable of. Similarly, the complexity of something like, for instance, a novel, or a feature film, eludes these technologies. The ways the tools become embedded, the ways the tools are treated or rejected, the ways they become social or cultural; that’s not for AI tools to do. That’s on us. Anyway, right now AI media is obsessed with its own nature and role in the world; it’s little better than a sequel to 2001: A Space Odyssey or Her. But like those films and countless other media objects, it does itself show us some of the ways we might either lean in to the change, or purposefully resist it. Any thoughts here on your own uses are very welcome!

    The creative and scientific methods blend in a fascinating way with AI media. Developers build tools that do a handful of things; users then learn to daisy-chain those tools together in personal workflows that suit their ideas and processes. To be truly innovative, creators will develop bold and strong original ideas (themes, stories, experiences), and then leverage their workflows to produce those ideas. It’s not just AI media. It’s AI media folded into everything else we already do, use, produce. That’s where the rubber meets the road, so to speak; where a tool or technique becomes the culture. That’s how it worked with printing and publishing, cinema and TV, computers, the internet, and that’s how it will work with AI. That’s where we’re headed. It’s not the singularity. It’s not the end of the world. it’s far more boring and fascinating than either of those could ever hope to be.

  • The roaring 2020s and their spectacular picture palaces

    Blank screen, auditorium to yourself, can’t lose. Photo by me, 18 April 2024.

    I took myself off to the movies lastnight. First time since 1917. The Sam Mendes film I mean, uh, obviously.

    Having gone on my little Godzilla binge earlier in the year, I thought it fitting that I take myself out to the latest instalment. The film itself was fine. Good loud dumb fun. Exactly the same formula as the others. A great soundtrack. Rebecca Hall being her wonderful earnest self. Dan Stevens being… whatever he is now (though he’ll always be Matthew to me). Content to one side, though, it was just great to be in the cinema again. For someone who allegedly studies the stuff from time to time, I don’t watch as much as I’d like; and I certainly don’t go to the cinema often at all. Lastnight showed me I probably should change that.

    I’ve often ruminated, in text and in brain, about the changing media landscape. I’m far from the only one, and recently Paris Marx put up a post about his quest to find Dune: Part One on home media. This story resonated with me. I have a sizeable physical media collection; it’s a dear asset and hobby, and one I am constantly surprised is not even close to mainstream nowadays.

    The production of physical has shifted considerably as demand has waned in the streaming era. DVDs are still, somehow, fairly popular; mostly due to an ageing and/or non-discerning audience (though that last bastion of DVDs, the supermarket DVD section, seems to have died off, finally). Blurays maintain a fair market share, but still require specialist hardware and are region-locked. Despite 4K Blurays being region-free and, with even a semi-decent TV, utterly gorgeous, they hold next to nothing of the market, being really only targeted at huge AV nerds like me.

    During COVID, the streaming platforms cemented their place in the homes and lives of everyone. I am certainly no exception to this. It was insanely convenient to have pretty much the world’s media output at the touch of a button. It was a good time: subscription prices were still relatively low, and the catalogues were decent enough to be worth having more than one or two services on the Apple TV at any given time.

    Netflix, Stan (an Aussie streaming service), and Prime Video were staples. They were also producing their own content, so in a way, they were modelling themselves on the megalithic studios of yore — as producers, distributors, marketers, even as archivists of popular culture.

    Things change, of course. They always do.

    Post-COVID, catalogues were culled. Most streaming services were operating, if not at a loss, then at least just breaking even with the equation of producing original content and/or buying distribution rights to older properties, or just other stuff in general.

    Then the original producers (in some cases the original studios) figured out they could just do it themselves. Disney+, Paramount+, Sony Core (aka Columbia); their own catalogues, their own archives, their own films straight from the cinema deal to the home media deal with no pesky negotiation.

    Prices for all streaming services have steadily risen over the last few years. For your average household, hell, even your above-average household, having all subscriptions active at one time simply isn’t feasible. It’s usually a question of who’s got what content at what time; or employ our house’s strategy and binge one or two platforms in one- or two-monthly bursts.

    Finding something specific in a given streaming catalogue is not a given. So you either pay Apple or Google or whoever to rent for a day or two or a week or whatever; or you pay them to ‘lease’ a copy of the film for you to view on-demand (they call this ‘buying’ the film). If giving money to the megacorps isn’t what you had in mind, maybe your brain would turn to the possibility of buying a physical copy of said media item for yourself.

    So you load up a web browser and punch in your best local media retailer. In my case, it’s a loud yellow behemoth called JB Hi-Fi; for more obscure titles or editions, it’d be something like Play DVD. These places are thin on the ground and, increasingly, even thin in the cloud. But JB’s physical media collection is dwindling, and has been for years. Their DVD/Bluray shelves used to occupy half of their huge stores; now they have maybe half a dozen tucked down the back, with the old real estate now occupied by more drones, more influencer starter kits, more appliances or pop culture paraphernalia.

    It struck me lastnight, as I headed into the cinema, that perhaps the film experience could see a bit of a bump if streaming services continue to fracture, and if physical media stock continues to disappear. If it’s a specific film that you want to see, and you know it’s on at the cinema, it’s probably more efficient overall to go and see it then and there. There are no guarantees any given film will be put up on a given streaming platform, nor that it will even get a physical media release any more. And if it does appear in either form, what quality will it be in? Would the experience be somehow diminished?

    There’s also something to be said for the sheer ubiquity and disposability of media in our current moment, particularly within the home, or home-based clouds. If I spot something on Netflix, I’ll add it to my List. I may watch it, but 7 times out of 10, I’ll forget it existed; once Netflix changes their catalogue, that item just floats away. I’m not notified; I’m not warned; unless it’s something on my watchlist on Letterboxd, or in a note or something, it just vanishes into the ether. Similarly with home media; if there’s a sale on at JB for Blurays, I might pick up a couple. They’ll then go on the shelf with the many, many others, and it might take years until I eventually get to it.

    There’s an intentionality to seeing a film at the cinema. In general, people are there to be absorbed in a singular experience. To not be distracted. To actually switch off from the outside world. I don’t claim any media superiority; I am a big old tech bro through and through, but there is something to the, ahem, CINEMATIC EXPERIENCE that really does retain the slightest touch of magic.

    So yes, perhaps we will see a little hike in moviegoing, if the platform economy continues to expand, explode, consume. Either that, or torrents will once again be IN for 2025 and beyond. Switch on that VPN and place your bets.

  • Blinded by machine visions

    A grainy, indistinct black and white image of a human figure wearing a suit and tie. The bright photo grain covers his eyes like a blindfold.
    Generated with Adobe Firefly, prompts by me.

    I threw around a quick response to this article on the socials this morning and, in particular, some of the reactions I was seeing. Here’s the money quote from photographer Annie Leibovitz, when asked about the effects of AI tools, generative AI technology, etc, on photography:

    “That doesn’t worry me at all,” she told AFP. “With each technological progress, there are hesitations and concerns. You just have to take the plunge and learn how to use it.”1

    The paraphrased quotes continue on the following lines:

    She says AI-generated images are no less authentic than photography.

    “Photography itself is not really real… I like to use PhotoShop. I use all the tools available.”

    Even deciding how to frame a shot implies “editing and control on some level,” she added.2

    A great many folx were posting responses akin to ‘Annie doesn’t count because she’s in the 1%’ or ‘she doesn’t count because she’s successful’, ‘she doesn’t have to worry anymore’ etc etc.

    On the one hand it’s typical reactionary stuff with which the socials are often ablaze. On the other hand, it’s fair to fear the impact of a given innovation on your livelihood or your passion.

    As I hint in my own posts3, though, I think the temptation to leap on this as privilege is premature, and a little symptomatic of whatever The Culture and/or The Discourse is at the moment, and has been for the duration of the platformed web, if not much longer.

    Leibovitz is and has always been a jobbing artist. Sure, in later years she has been able to pick and choose a little more, but by all accounts she is a busy and determined professional, treating every job with just as much time, effort, dedication as she always has. The work, for Leibovitz, has value, just as much — if not more — than the product or the paycheck.

    I don’t mean to suddenly act my age, or appear much older and grumpier than I am, but I do wonder about how much time aspiring or current photographers spend online discussing and/or worrying and/or reacting to the latest update or the current fad-of-the-moment. I 100% understand the need for today’s artists and creators to engage in some way with the social web, if only to put their names out there to try and secure work. But if you’re living in the comments, whipping yourselves and others into a frenzy about AI or whatever it is, is that really the best use of your time?

    The irony of me asking such questions on a blog where I do nothing but post and react is not lost on me, but this blog for me is a scratchpad, a testing ground, a commonplace book; it’s a core part of my ‘process’, whatever that is, and whatever it’s for. This is practice for other writing, for future writing, for my identity, career, creative endeavours as a writer. It’s a safe space; I’m not getting angry (necessarily), or seeking out things to be angry about.

    But I digress. Leibovitz is not scared of AI. And as someone currently working in this space, I can’t disagree. Having even a rudimentary understanding of what these tools are actually doing will dispel some of the fear.

    Further, photography, like the cinema that it birthed, has already died a thousand deaths, and will die a thousand more.

    Brilliant4 photography lecturer and scholar Alison Bennett speaks to the legacy and persistence of photographic practice here:

    “Recent examples [of pivotal moments of change in photography] include the transition from analogue film to digital media in the late 20th century, then the introduction of the internet-connected smart phone from 2007,” they said.

    “These changes fundamentally redefined what was possible and how photography was used.

    “The AI tipping point is just another example of how photography is constantly being redefined.”5

    As ever, the tools are not the problem. The real enemies are the companies and people that are driving the tools into the mainstream at scale. The companies that train their models on unlicensed datasets, drawn from copyrighted material. The people that buy into their own bullshit about AI and AGI being some kind of evolutionary and/or quasi-biblical moment.

    For every post shitting on Annie Leibovitz, you must have at least twenty posts actively shitting on OpenAI and their ilk, pushing for ethically-sourced and maintained datasets, pushing for systemic change to the resource management of AI systems, including sustainable data centers.

    The larger conceptual questions are around authenticity and around hard work. If you use AI tools, are you still an authentic artist? Aren’t AI tools just a shortcut? Of course, the answers are ‘not necessarily’. If you’ve still done the hard yards to learn about your craft, to learn about how you work, to discover what kinds of stories and experiences you want to create, to find your voice, in whatever form it takes, then generative AI is a paintbrush. A weird-looking paintbrush, but a paintbrush nevertheless (or plasticine, or canvas, or glitter, or an app, etc. etc. ad infinitum).

    Do the work, and you too can be either as ambivalent as Leibovitz, or as surprised and delighted as you want to be. Either way, you’re still in control.

    Notes ↩︎

    1. Agence France-Presse 2024, ‘Photographer Annie Leibovitz: “AI doesn’t worry me at all”’, France 24, viewed 26 March 2024, <https://www.france24.com/en/live-news/20240320-photographer-annie-leibovitz-ai-doesn-t-worry-me-at-all>.
      ↩︎
    2. ibid. ↩︎
    3. See here, and with tiny edits for platform affordances here and here. What’s the opposite of POSSE? PEPOS? ↩︎
    4. I am somewhat biased as, at the time of writing, Dr. Bennett and I currently share a place of work. To look through their expanded (heh) works, go here. ↩︎
    5. Odell, T 2024, ‘New exhibition explores AI’s influence on the future of photography’, RMIT University, viewed 26 March 2024, <https://www.rmit.edu.au/news/all-news/2024/mar/photo-2024>.
      ↩︎
  • Godzilla: King of the Franchises

    If you like what’s going on here at The Clockwork Penguin, if you appreciate the cut of my particular jib, as it were, buy me a coffee!

    A film still from Godzilla vs Kong (2021), featuring Godzilla filling the frame from the torso up. It is raining heavily, so we can't see much detail of Godzilla's face, but we can clearly see blue lightning running along the spines on his back.
    Godzilla with the spicy lightning as depicted in Godzilla vs. Kong (2021).

    The 2014 reboot/continuation/expansion of the Godzilla franchise opens with the standard mystery box. A helicopter flies low over a jungle landscape, there are low minor chords from a rumbling orchestra: dissonance, uncertainty, menace. Helicopter Passenger #1 turns to Helicopter Passenger #2: “They found something.”

    This is the germ of what is now a multi-film franchise, with a spin-off TV series that debuted in late 2023. A few weeks ago, I re-watched Gareth Edwards’ 2014 reboot, as well as the sequel films I hadn’t seen, Godzilla II: King of the Monsters and Godzilla vs. Kong.

    It was a bit of fun, obviously, a last hurrah before I went back to the very serious business of media academicking, but as is wont to happen, it’s been stewing ever since. So here: have some little thoughts on big monsters.


    Last week I Masta’d1 up some speculations as to why Argylle has flopped. The first and most obvious reason that a film might tank is that it’s just not a very good film, as this may well be true of Argylle. But in a time where cinema is dead and buried and a media object is never discrete, we can’t look at the film in a vacuum.

    I have thoughts on why #Argylle flopped. I haven’t seen it, so I won’t go into any great depth, but suffice to say there are two major components:

    1) Marvel killed transmedia storytelling, jumped around on the corpse, drove a steamroller over it, then buried it in a nuclear waste facility.

    2) Camp doesn’t hit like it used to. Big ensemble campy treats aren’t as sweet now; in an age of hyper-sensitivity, broad knowledge and information access, they taste a little sour. Ain’t no subtext anymore.2

    The marketing machine behind Argylle decided they’d play a little game, by teasing a novel written by a mystery author (both in terms of them not being well-known, but also an author of actual mystery), with the film being quickly picked up for production. This was fairly clumsily-done, but leaving that to one side: okay, cool idea. The conceit is that the author runs into the real-life equivalent of one of their characters who whisks them away on an adventure. Cue ideas of unreliable narration, possible brainwashing, or whatever, and there’s the neat little package.

    The concept overall is solid, but Universal and Apple made the mistake of thinking they could shoehorn this concept into a campaign that ‘tricked’ the audience into thinking some of it was factual, or at least had some tenuous crossover with reality.

    Basically, they tried an old-school transmedia campaign.

    Transmedia storytelling has always been around in some form or another. It dovetails quite nicely with epistolary and experimental narratives, like Mary Shelley’s original Frankenstein, and discussions of transmedia also work well when you’re thinking about serial stories or adaptations. The term is most often attributed to Henry Jenkins, a wonderful and kindly elder scholar who was thinking about the huge convergence of media technologies occasioned by the wide adoption of the internet in the late 1990s and early 2000s.

    Jenkins’ ur-example of transmedia is The Matrix franchise, “a narrative so large that it cannot be contained within a single medium.”3 The idea is that in order to truly appreciate the narrative as a whole, the audience has to follow up on all the elements, be they films, video games, comic books, or whatever.

    “Each franchise entry needs to be self-contained so you don’t need to have seen the film to enjoy the game, and vice versa. Any given product is a point of entry into the franchise as a whole. Reading across the media sustains a depth of experience that motivates more consumption.”4

    This model is now far from unique in terms of marketing or storytelling; with the MCU, DC Universe, the Disneyfied Star Wars universe and others, we have no dearth of complex narratives and storyworlds to choose from. This is maybe now partly why transmedia is seen as, at best, a little dated, old hat, and at worst, a bit of a dirty word, when it comes to narrative, media, or cinema studies. Those still chipping away at the transmedia stoneface are seen as living in the past or worse, wasting their time. I don’t think it’s a waste of time, nor do I necessarily see it as living in the past; it’s just that transmedia is media now.

    Every new media commodity, be it a film, an app, a game, a platform, novel, has a web of attendant media commodities spring up around it. Mostly these are used for marketing, but occasionally these extraneous texts may relay some plot point or narrative element. The issue is that you need to conceit to be front and centre, you need some idea of the full narrative; you can’t expect the audience to want to do anything. The producers of Argylle made this mistake. They did transmedia the old-fashioned way, where narrative elements are spread across discrete media objects (e.g. book and film), and they expected the audience to want to fill in the gaps, and to share their excitement at having done so… on social media, I guess?

    But like transmedia storytelling, social media ain’t what she used to be. Our present internet is fragmented, hyper-platformed, paywalled; city-states competing for dominance, for annexation (acquisition), for citizens or slaves (subscribers). Content is still king, but the patrician’s coffers care not as to whether that content is produced by the finest scribes of the age, or the merchant guild’s new automatons.

    Viral is still viral, popular is still popular, but the propagation of content moves differently. Hashtags, likes, views don’t mean much anymore. You want people talking, but you only care as much as it gets new people to join your platform or your service. Get new citizens inside the gates, then lock and bar the gates behind them; go back to the drawing board for the next big campaign. The long tail is no more; what matters is flash in the pan attention attacks.


    The producers behind the Godzilla reboot clearly envisioned a franchise. This is clear enough from how the film ends (or more accurately, where it stops talking, and the credits are permitted to roll). Godzilla apparently saves the world (or at least Honolulu and Las Vegas) from another giant monster, then disappears into the sea, leaving humanity to speculate as to its motivations.

    It’s also apparent that the filmmakers didn’t want a clean break from the cultural history, themes or value of the broader Godzilla oeuvre; the title sequence suggests that the 1954 Castle Bravo tests were actually an operation to destroy Godzilla5. And in the film’s prologue, this wonderful shot is presented with virtually no context.

    A film still from Godzilla (2014), with a Lego model of the Saturn V rocket in the foreground, and a poster of Godzilla in the background.
    American triumphalism meets Japanese… er, monster promotions?

    What struck me most, though, is the lack of overt story-bridges, particularly in the first film. Story-bridges are parts of the plot, e.g. characters, events, images, that allow the audience to jump off to another part of the narrative. These jumping-off points can be explicit, e.g. an end-credits sequence, or a line of dialogue referring to a past/future event, or they can be implied, e.g. the introduction of a character in a minor role that may participate more prominently in other media.

    As media franchises become more complex, these points/bridges are not as often modelled as connecting branches to nodes around a centred point (a tentpole film, for instance), but as a mesh that connects multiple, interconnected tentpole media. In some of my academic work, with my colleagues Vashanth Selvadurai and Peter Vistisen, we’ve explored how Marvel attempts to balance this complexity:

    “[Marvel] carefully balances production by generating self-contained stories for the mass audience, which include moments of interconnectivity in some scenes to fortify the MCU and thereby accommodate the fan base… [T]he gradual scaling up in bridge complexity from characters to storyworld to a cohesive storyline is integrated into a polycentric universe of individual transmedia products. These elements are not gathered around one tentpole event that the audience has to experience before being able to make sense of the rest.”4

    In Godzilla, the story-bridges are more thematic, even tonal. The story remains consistently about humanity’s desire to control the natural world, and that world’s steadfast resistance to control; multiple times we hear the scientist Ishiro Serizawa (Ken Watanabe) speak about ‘balance’ in nature, the idea of a natural equilibrium that humanity has upset, and that Godzilla and his kin will restore. There is also a balance within and between humanity; corporate greed, political power struggles, individual freedoms and restrictions, all vie to find a similar kind of equilibrium, if such a thing is possible. The resulting tone is one that feels universal but prescient; topical and relevant to the contemporary moment, despite the presence of enormous monsters.

    This tone is carried over into Godzilla: King of the Monsters. The monsters multiply in this instalment, and an extraterrestrial element is introduced, but in general Godzilla’s animal motivation is about preservation of self, but also of Earth and its biology. I should also note that there are more explicit bridges in this film, like the characters of Dr. Serizawa and his colleague Dr. Vivienne Graham (Sally Hawkins). But the true connecting thread, at least for me, is this repeated theme of humans and our puny power struggles playing out against the backdrop of a deep time, a history, forces and powers so ancient that we can never really understand them.

    This macro-bridge, if you like, allows the filmmakers to then make tweaks to the micro-elements of the story. If they want or need to adjust the character focus, they can. If the plot of a single film seems a little rote, maybe, or they want to try something different, they’ve given themselves enough space in the story and the story-world to do that. This may not necessarily be intentional, but it certainly appears as an effective counter-model to the MCU/Disney mode, where everything seems over-complicated and planned out in multi-year phases, and everything is so locked in. The MonsterVerse approach is one of ad hoc franchise storytelling, and the result is a universe that feels more free, more open: full of possibilities and untold stories.

    The point of all of this, I suppose, is to let us see what works and what doesn’t. As a storyteller or creative type, it helps me to model and test approaches to storytelling of all scales and budgets, as I think about what kinds of narratives I want to develop, and in which media form. Beyond that, though, I think that as we move into a contentscape that muddles the human-made with the computer-generated, this kind of analysis and discussion is more essential than ever.


    Notes & References

    1. Still working out the vernacular for the new social web. ↩︎
    2. Me, on Mastodon, February 6, 2024. ↩︎
    3. Jenkins, H. (2006). Convergence culture: Where old and new media collide. New York University Press, p. 95. ↩︎
    4. Jenkins, Convergence culture, p. 96. ↩︎
    5. The first Godzilla film, directed by Ishirō Honda, was also released in 1954 ↩︎
    6. Selvadurai, V., Vistisen, P., & Binns, D. (2022). Bridge Complexity as a Factor in Audience Interaction in Transmedia Storytelling. Journal of Asia-Pacific Pop Culture, 7(1), 85–108 (quote from pages 96-7). https://doi.org/10.5325/jasiapacipopcult.7.1.0085 ↩︎
  • Critics and creation

    Photo by Leah Newhouse on Pexels.

    I started reading this interview this morning, between Anne Helen Peterson and Betsy Gaines Quammen. I still haven’t finished reading, despite being utterly fascinated, but even before I got to the guts of the interview, I was struck by a thought:

    In the algorithmised world, the creator is the critic.

    This thought is not necessarily happening in isolation; I’ve been thinking about ‘algorithmic culture’ for a couple of years, trying to order these thoughts into academic writing, or even creative writing. But this thought feels like a step in the right direction, even if I’ve no idea what the final output should or will be. Let’s scribble out some notes…

    If there’s someone whose work we enjoy, they’ll probably have an online presence — a blog or social media feed we can follow — where they’ll share what they like.

    It’s an organic kind of culture — but it’s one where the art and vocation of the critic continues to be minimised.

    This — and associated phenomena — is the subject of a whole bunch of recent and upcoming books (including this one, which is at the top of my to-read pile for the next month): a kind of culture where the all-powerful algorithm becomes the sole arbiter of taste, but I also think there is pressure on creatives to be their own kind of critical and cultural hub.

    On the inverse, what we may traditionally have called critics — so modern-day social media commentators, influencers, your Booktubers or Booktokkers, your video essayists and their ilk — now also feel pressure to create. This pressure will come from their followers and acolytes, but also from random people who encounter them online, who will say something like “if you know so much why don’t you just do it yourself” etc etc…

    Some critics will leap at the opportunity and they absolutely should — we are hearing from diverse voices that wouldn’t otherwise have thought to try.

    But some should leave the creation to others — not because they’re not worth hearing from, they absolutely are — but because their value, their creativity, their strength, lies in how they shape language, images, metaphor, around the work of others. They don’t realise — as I didn’t for a long time — that being a critic is a vocation, a life’s work, a real skill. Look at any longer-form piece in the London Review of Books or The New Inquiry and it becomes very clear how valuable this work is.

    I’ve always loved the term critic, particularly cultural critic, or commentator, or essayist… they always seemed like wonderful archaic terms that don’t belong in the modern, fragmented, divided, confused world. But to call oneself a critic or essayist, to own that, and only that, is to defy the norms of culture; to refuse the ‘pillars’ of novel, film, press/journalism, and to stand to one side, giving much-needed perspective to how these archaic forms define, reflect, and challenge society.