Stubsack: weekly thread for sneers not worth an entire post, week ending Sunday 27 October 2024
  • "Initials" by "Florian Körner", licensed under "CC0 1.0". / Remix of the original. - Created with dicebear.comInitialsFlorian Körnerhttps://github.com/dicebear/dicebearBL
    BlueMonday1984
    3h ago 100%

    Now that the content mafia has realized GenAI isn’t gonna let them get rid of all the expensive and troublesome human talent. it’s time to give Big AI a wedgie.

    Considering the massive(ly inflated) valuations running around Big AI and the massive amounts of stolen work that powers the likes of CrAIyon, ChatGPT, DALL-E and others, I suspect the content mafia is likely gonna try and squeeze every last red cent they can out of the AI industry.

    6
  • Stubsack: weekly thread for sneers not worth an entire post, week ending Sunday 27 October 2024
  • "Initials" by "Florian Körner", licensed under "CC0 1.0". / Remix of the original. - Created with dicebear.comInitialsFlorian Körnerhttps://github.com/dicebear/dicebearBL
    BlueMonday1984
    9h ago 100%

    Considering Glaze and Nightshade have been around for a while, and I talked about sabotaging scrapers back in July, arguably, it already has.

    Hell, I ran across a much smaller scale case of this a couple days ago:

    Not sure how effective it is, but if Elon's stealing your data for his autoplag no matter what, you might as well try to force-feed it as much poison as you can.

    7
  • Sam Altman rebrands Worldcoin to ‘World’, still wants your eyeballs
  • "Initials" by "Florian Körner", licensed under "CC0 1.0". / Remix of the original. - Created with dicebear.comInitialsFlorian Körnerhttps://github.com/dicebear/dicebearBL
    BlueMonday1984
    21h ago 100%

    I’m trying to think of how you monetize eyeball scans and the first thing that comes to mind (well, after being able to break biometric security) is training an AI to generate fake but passable eyeballs to undercut the use of iris scans as an anti-bot tool.

    Silicon Valley's basically an AI cult at this point, so I can see your case.

    6
  • Stubsack: weekly thread for sneers not worth an entire post, week ending Sunday 20 October 2024
  • "Initials" by "Florian Körner", licensed under "CC0 1.0". / Remix of the original. - Created with dicebear.comInitialsFlorian Körnerhttps://github.com/dicebear/dicebearBL
    BlueMonday1984
    22h ago 100%

    New piece from The Atlantic: The Age of AI Child Abuse is Here, which delves into a large-scale hack of Muah.AI and the large-scale problem of people using AI as a child porn generator.

    And now, another personal sidenote, because I cannot stop writing these (this one's thankfully unrelated to the article's main point):

    The idea that "[Insert New Tech] Is Inevitable^tm^" (which Unserious Academic interrogated in depth BTW) took a major blow when NFTs crashed and burned in full view of the public eye and got rapidly turned into a pop-culture punchline.

    That, I suspect, is helping to fuel the large scale rejection of AI and resistance to its implementation - Silicon Valley's failure to make NFTs a thing has taught people that Silicon Valley can be beaten, that resistance is anything but futile.

    8
  • Stubsack: weekly thread for sneers not worth an entire post, week ending Sunday 20 October 2024
  • "Initials" by "Florian Körner", licensed under "CC0 1.0". / Remix of the original. - Created with dicebear.comInitialsFlorian Körnerhttps://github.com/dicebear/dicebearBL
    BlueMonday1984
    3d ago 100%

    anyone wanna take bets on how much pearlclutching surprisedpikachu we’ll see

    I suspect we'll see a fair amount. Giving some specifics:

    1. I suspect we'll see Sammy accused of endangering all of humanity for a quick buck - taking Altman at his word, OpenAI is attempting to create something which they themselves believe could wipe out humanity if they screw things up.

    2. I expect calls to regulate the AI industry will louden in response to this - what Sammy's doing here is giving the true believers more ammo to argue Silicon Valley may potentially trigger the robot apocalypse that Silicon Valley themselves have claimed AI is capable of unleashing.

    10
  • Stubsack: weekly thread for sneers not worth an entire post, week ending Sunday 20 October 2024
  • "Initials" by "Florian Körner", licensed under "CC0 1.0". / Remix of the original. - Created with dicebear.comInitialsFlorian Körnerhttps://github.com/dicebear/dicebearBL
    BlueMonday1984
    4d ago 100%

    New piece from The Atlantic: The AI Boom Has an Expiration Date

    The full piece is worth a read, but the conclusion's pretty damn good, so I'm copy-pasting it here:

    All of this financial and technological speculation has, however, created something a bit more solid: self-imposed deadlines. In 2026, 2030, or a few thousand days, it will be time to check in with all the AI messiahs. Generative AI—boom or bubble—finally has an expiration date.

    9
  • Google, Amazon buy nonexistent mini nuclear reactors for AI data centers
  • "Initials" by "Florian Körner", licensed under "CC0 1.0". / Remix of the original. - Created with dicebear.comInitialsFlorian Körnerhttps://github.com/dicebear/dicebearBL
    BlueMonday1984
    4d ago 100%

    If these nuclear plants manage to come to fruition, it'll be the sole miniscule silver lining of the bubble. Considering its AI, though, I expect they'll probably suffer some kind of horrific Chernobyl-grade accident which kills nuclear power for good, because we can't have nice things when there's AI involved.

    21
  • Stubsack: weekly thread for sneers not worth an entire post, week ending Sunday 20 October 2024
  • "Initials" by "Florian Körner", licensed under "CC0 1.0". / Remix of the original. - Created with dicebear.comInitialsFlorian Körnerhttps://github.com/dicebear/dicebearBL
    BlueMonday1984
    5d ago 100%

    the lasting legacy of GenAI will be a elevated background level of crud and untruth, an erosion of trust in media in general, and less free quality stuff being available.

    I personally anticipate this will be the lasting legacy of AI as a whole - everything that you mentioned was caused in the alleged pursuit of AGI/Superintelligence^tm^, and gen-AI has been more-or-less the "face" of AI throughout this whole bubble.

    I've also got an inkling (which I turned into a lengthy post) that the AI bubble will destroy artificial intelligence as a concept - a lasting legacy of "crud and untruth" as you put it could easily birth a widespread view of AI as inherently incapable of distinguishing truth from lies.

    9
  • Stubsack: weekly thread for sneers not worth an entire post, week ending Sunday 20 October 2024
  • "Initials" by "Florian Körner", licensed under "CC0 1.0". / Remix of the original. - Created with dicebear.comInitialsFlorian Körnerhttps://github.com/dicebear/dicebearBL
    BlueMonday1984
    5d ago 100%

    It was a pretty good comment, and pointed out one of the possible risks this AI bubble can unleash.

    I've already touched on this topic, but it seems possible (if not likely) that copyright law will be tightened in response to the large-scale theft performed by OpenAI et al. to feed their LLMs, with both of us suspecting fair use will likely take a pounding. As you pointed out, the exploitation of fair use's research exception makes it especially vulnerable to its repeal.

    On a different note, I suspect FOSS licenses (Creative Commons, GPL, etcetera) will suffer a major decline in popularity thanks to the large-scale code theft this AI bubble brought - after two-ish years of the AI industry (if not tech in general) treating anything publicly available as theirs to steal (whether implicitly or explicitly), I'd expect people are gonna be a lot stingier about providing source code or contributing to FOSS.

    11
  • Stubsack: weekly thread for sneers not worth an entire post, week ending Sunday 20 October 2024
  • "Initials" by "Florian Körner", licensed under "CC0 1.0". / Remix of the original. - Created with dicebear.comInitialsFlorian Körnerhttps://github.com/dicebear/dicebearBL
    BlueMonday1984
    5d ago 100%

    The top comment's also pretty good, especially the final paragraph:

    I guess these companies decided that strip-mining the commons was an acceptable deal because they’d soon be generating their own facts via AGI, but that hasn’t come to pass yet. Instead they’ve pissed off many of the people they were relying on to continue feeding facts and creativity into the maws of their GPUs, as well as possibly fatally crippling the concept of fair use if future court cases go against them.

    15
  • Stubsack: weekly thread for sneers not worth an entire post, week ending Sunday 20 October 2024
  • "Initials" by "Florian Körner", licensed under "CC0 1.0". / Remix of the original. - Created with dicebear.comInitialsFlorian Körnerhttps://github.com/dicebear/dicebearBL
    BlueMonday1984
    7d ago 100%

    Zitron's given commentary on PC Gamer's publicly pilloried pro-autoplag piece:

    He's also just dropped a thorough teardown of the tech press for their role in enabling Silicon Valley's worst excesses. I don't have a fitting Kendrick Lamar reference for this, but I do know a good companion piece: Devs and the Culture of Tech, which goes into the systemic flaws in tech culture which enable this shit.

    15
  • www.wheresyoured.at

    Gonna add the opening quote, because it is glorious: > You cannot make friends with the rock stars...if you're going to be a true journalist, you know, a rock journalist. First, you never get paid much, but you will get free records from the record company. > > [There’s] fuckin’ nothin' about you that is controversial. God, it's gonna get ugly. And they're gonna buy you drinks, you're gonna meet girls, they're gonna try to fly you places for free, offer you drugs. I know, it sounds great, but these people are not your friends. You know, these are people who want you to write sanctimonious stories about the genius of the rock stars and they will ruin rock 'n' roll and strangle everything we love about it. > > Because they're trying to buy respectability for a form that's gloriously and righteously dumb. > > Lester Bangs, Almost Famous (2000) EDITED TO ADD: If you want a good companion piece to this, [Devs and the Culture of Tech](https://unserious.substack.com/p/devs-pt1) by @UnseriousAcademic is a damn good read, going deep into the cultural issues which leads to the fawning tech press Zitron so thoroughly tears into.

    26
    3

    Need to let loose a primal scream without collecting footnotes first? Have a sneer percolating in your system but not enough time/energy to make a whole post about it? Go forth and be mid: Welcome to the Stubsack, your first port of call for learning fresh Awful you’ll near-instantly regret. Any awful.systems sub may be subsneered in this subthread, techtakes or no. If your sneer seems higher quality than you thought, feel free to cut’n’paste it into its own post — there’s no quota for posting and the bar really isn’t that high. > The post Xitter web has spawned soo many “esoteric” right wing freaks, but there’s no appropriate sneer-space for them. I’m talking redscare-ish, reality challenged “culture critics” who write about everything but understand nothing. I’m talking about reply-guys who make the same 6 tweets about the same 3 subjects. They’re inescapable at this point, yet I don’t see them mocked (as much as they should be) > > Like, there was one dude a while back who insisted that women couldn’t be surgeons because they didn’t believe in the moon or in stars? I think each and every one of these guys is uniquely fucked up and if I can’t escape them, I would love to sneer at them. [Last week's thread](https://awful.systems/post/2559278) (Semi-obligatory thanks to @dgerard for [starting this](https://awful.systems/post/1162442))

    18
    245
    https://awful.systems/comment/4979228

    (This is basically an expanded version of a comment on the weekly Stubsack - I've linked it above for convenience's sake.) This is pure gut instinct, but I’m starting to get the feeling this AI bubble’s gonna destroy the concept of artificial intelligence as we know it. On the artistic front, there's the general [tidal wave of AI-generated slop](https://nymag.com/intelligencer/article/ai-generated-content-internet-online-slop-spam.html) (which I've come to term "the slop-nami") which has come to drown the Internet in zero-effort garbage, interesting only when the art's [utterly insane](https://twitter.com/FacebookAIslop) or its prompter gets [publicly humiliated](https://twitter.com/venturetwins/status/1844127816612499907), and, to quote Line Goes Up, "derivative, lazy, ugly, hollow, and boring" the other 99% of the time. (And all while the AI industry [steals artists' work](https://spectrum.ieee.org/midjourney-copyright), [destroys their livelihoods](https://archive.is/GGNoC) and [shamelessly mocks their victims throughout](https://twitter.com/tsarnick/status/1803920566761722166).) On the "intelligence" front, the bubble's given us public and spectacular failures of reasoning/logic like Google [gluing pizza](https://www.theverge.com/2024/6/11/24176490/mm-delicious-glue) and [eating onions](https://www.avclub.com/google-s-ai-feeds-answers-from-the-onion-1851500362), ChatGPT [sucking at chess](https://www.chess.com/article/view/chatgpt-chess-advice) and [briefly losing its shit](https://arstechnica.com/information-technology/2024/02/chatgpt-alarms-users-by-spitting-out-shakespearean-nonsense-and-rambling/), and so much more - even in the absence of [formal proof LLMs can't reason](https://garymarcus.substack.com/p/llms-dont-do-formal-reasoning-and), its not hard to conclude they're far from intelligent. All of this is, of course, happening whilst the tech industry as a whole is hyping the ever-loving FUCK out of AI, breathlessly praising its supposed creativity/intelligence/brilliance and relentlessly claiming that they're on the cusp of AGI/superintelligence/whatever-the-fuck-they're-calling-it-right-now, they just need to raise a few more billion dollars and [boil a few more hundred lakes](https://www.standard.co.uk/news/tech/ai-chatgpt-water-power-usage-b1106592.html) and [kill a few more hundred species](https://pivot-to-ai.com/2024/07/23/data-centers-risk-missing-us-climate-goals-especially-with-ai/) and enable a few more months of [SEO](https://wired.me/gear/google-search-ai-spam/) and [scams](https://www.abc.net.au/news/2023-04-12/artificial-intelligence-ai-scams-voice-cloning-phishing-chatgpt/102064086) and [spam](https://www.businessinsider.com/ai-spam-google-ruin-internet-search-scams-chatgpt-2024-1) and [slop](https://futurism.com/the-byte/source-ai-slop-facebook) and soulless shameless scum-sucking shitbags senselessly shitting over everything that was good about the Internet. ---- The public's collective consciousness was ready for a lot of futures regarding AI - a future where it took everyone's jobs, a future where it started the apocalypse, a future where it brought about utopia, etcetera. A future where AI ruins everything by being utterly, fundamentally incompetent, like the one we're living in now? That's a future the public was *not* ready for - sci-fi writers weren't playing much the idea of "incompetent AI ruins everything" (*[Paranoia](https://en.wikipedia.org/wiki/Paranoia_(role-playing_game))* is the only example I know of), and the tech press wasn't gonna run stories about AI's faults until it became unignorable (like that lawyer who [got in trouble for taking ChatGPT at its word](https://arstechnica.com/tech-policy/2023/05/lawyer-cited-6-fake-cases-made-up-by-chatgpt-judge-calls-it-unprecedented/)). Now, of course, the public's had plenty of time to let the reality of this current AI bubble sink in, to watch as the AI industry tries and fails to fix the unfixable hallucination issue, to watch the likes of CrAIyon and Midjourney continually fail to produce anything even remotely worth the effort of typing out a prompt, to watch AI creep into and enshittify every waking aspect of their lives as their bosses and higher-ups buy the hype hook, line and fucking sinker. ---- All this, I feel, has built an image of AI as inherently incapable of humanlike intelligence/creativity (let alone Superintelligence^tm^), no matter how many server farms you build or oceans of water you boil. Especially so on the creativity front - publicly rejecting AI, like what [Procreate](https://twitter.com/Procreate/status/1825311104584802470) and [Schoolism](https://twitter.com/SchoolismLIVE/status/1843796461638824034) did, earns you an instant standing ovation, whilst openly shilling it (like [PC Gamer](https://twitter.com/pcgamer/status/1844796247044948363) or [The Bookseller](https://twitter.com/thebookseller/status/1841828201804300491)) or showcasing it (like [Justine Moore](https://twitter.com/venturetwins/status/1844127816612499907), [Proper Prompter](https://twitter.com/ProperPrompter/status/1809206091764613487) or [Luma Labs](https://twitter.com/LumaLabsAI/status/1800921393321934915)) gets you publicly and relentlessly lambasted. To [quote Baldur Bjarnason](https://www.baldurbjarnason.com/2024/slop-framing-failure-as-success/), the “E-number additive, but for creative work” connotation of “AI” is more-or-less a permanent fixture in the public’s mind. I don't have any pithy quote to wrap this up, but to take a shot in the dark, I expect we're gonna see a particularly long and harsh AI winter once the bubble bursts - one fueled not only by disappointment in the failures of LLMs, but widespread public outrage at the massive damage the bubble inflicted, with AI funding facing heavy scrutiny as the public comes to treat any research into the field as done with potentally malicious intent.

    16
    2

    Need to let loose a primal scream without collecting footnotes first? Have a sneer percolating in your system but not enough time/energy to make a whole post about it? Go forth and be mid: Welcome to the Stubsack, your first port of call for learning fresh Awful you’ll near-instantly regret. Any awful.systems sub may be subsneered in this subthread, techtakes or no. If your sneer seems higher quality than you thought, feel free to cut’n’paste it into its own post — there’s no quota for posting and the bar really isn’t that high. > The post Xitter web has spawned soo many “esoteric” right wing freaks, but there’s no appropriate sneer-space for them. I’m talking redscare-ish, reality challenged “culture critics” who write about everything but understand nothing. I’m talking about reply-guys who make the same 6 tweets about the same 3 subjects. They’re inescapable at this point, yet I don’t see them mocked (as much as they should be) > > Like, there was one dude a while back who insisted that women couldn’t be surgeons because they didn’t believe in the moon or in stars? I think each and every one of these guys is uniquely fucked up and if I can’t escape them, I would love to sneer at them. [Last week’s thread](https://awful.systems/post/2505486) (Semi-obligatory thanks to @dgerard for [starting this](https://awful.systems/post/1162442))

    21
    169
    www.wheresyoured.at

    > None of what I write in this newsletter is about sowing doubt or "hating," but a sober evaluation of where we are today and where we may end up on the current path. I believe that the artificial intelligence boom — which would be better described as a generative AI boom — is (as I've said before) unsustainable, and will ultimately collapse. I also fear that said collapse could be ruinous to big tech, deeply damaging to the startup ecosystem, and will further sour public support for the tech industry. Can't blame Zitron for being pretty downbeat in this - given the AI bubble's size and side-effects, its easy to see how its bursting can have some cataclysmic effects. (Shameless self-promo: [I ended up writing a bit about the potential aftermath as well](https://awful.systems/post/2031653))

    70
    32

    Need to let loose a primal scream without collecting footnotes first? Have a sneer percolating in your system but not enough time/energy to make a whole post about it? Go forth and be mid: Welcome to the Stubsack, your first port of call for learning fresh Awful you’ll near-instantly regret. Any awful.systems sub may be subsneered in this subthread, techtakes or no. If your sneer seems higher quality than you thought, feel free to cut’n’paste it into its own post — there’s no quota for posting and the bar really isn’t that high. > The post Xitter web has spawned soo many “esoteric” right wing freaks, but there’s no appropriate sneer-space for them. I’m talking redscare-ish, reality challenged “culture critics” who write about everything but understand nothing. I’m talking about reply-guys who make the same 6 tweets about the same 3 subjects. They’re inescapable at this point, yet I don’t see them mocked (as much as they should be) > > Like, there was one dude a while back who insisted that women couldn’t be surgeons because they didn’t believe in the moon or in stars? I think each and every one of these guys is uniquely fucked up and if I can’t escape them, I would love to sneer at them. (Semi-obligatory thanks to @dgerard for [starting this](https://awful.systems/post/1162442))

    23
    245
    aeon.co

    This started as a summary of a random essay Robert Epstein (fuck, that's an unfortunate surname) cooked up back in 2016, and evolved into a diatribe about how the AI bubble affects how we think of human cognition. This is probably a bit outside awful's wheelhouse, but hey, this is MoreWrite. **The TL;DR** The general article concerns two major metaphors for human intelligence: * The information processing (IP) metaphor, which views the brain as some form of computer (implicitly a classical one, though you could probably cram a quantum computer into that metaphor too) * The anti-representational metaphor, which views the brain as a living organism, which constantly changes in response to experiences and stimuli, and which contains jack shit in the way of any computer-like components (memory, processors, algorithms, etcetera) Epstein's general view is, if the title didn't tip you off, firmly on the anti-rep metaphor's side, dismissing IP as "not even slightly valid" and openly arguing for dumping it straight into the dustbin of history. His main major piece of evidence for this is a basic experiment, where he has a student draw two images of dollar bills - one from memory, and one with a real dollar bill as reference - and compare the two. Unsurprisingly, the image made with a reference blows the image from memory out of the water every time, which Epstein uses to argue against any notion of the image of a dollar bill (or anything else, for that matter) being stored in one's brain like data in a hard drive. Instead, he argues that the student making the image had re-experienced seeing the bill when drawing it from memory, with their ability to do so having come because their brain had changed at the sight of many a dollar bill up to this point to enable them to do it. Another piece of evidence he brings up is a [1995 paper from *Science*](https://pubmed.ncbi.nlm.nih.gov/7725104/) by Michael McBeath regarding baseballers catching fly balls. Where the IP metaphor reportedly suggests the player roughly calculates the ball's flight path with estimates of several variables ("the force of the impact, the angle of the trajectory, that kind of thing"), the anti-rep metaphor (given by McBeath) simply suggests the player catches them by moving in a manner which keeps the ball, home plate and the surroundings in a constant visual relationship with each other. The final piece I could glean from this is [a report in *Scientific American*](https://www.scientificamerican.com/article/why-the-human-brain-project-went-wrong-and-how-to-fix-it/) about the Human Brain Project (HBP), a $1.3 billion project launched by the EU in 2013, made with the goal of simulating the entire human brain on a supercomputer. Said project went on to become a "brain wreck" less than two years in (and eight years before its 2023 deadline) - a "brain wreck" Epstein implicitly blames on the whole thing being guided by the IP metaphor. Said "brain wreck" is a good place to cap this section off - the essay is something I recommend reading for yourself (even if I do feel its arguments aren't particularly strong), and its not really the main focus of this little ramblefest. Anyways, onto my personal thoughts. **Some Personal Thoughts** Personally, I suspect the AI bubble's made the public a lot less receptive to the IP metaphor these days, for a few reasons: 1) Articial Idiocy The entire bubble was *sold* as a path to computers with human-like, if not godlike intelligence - artificial thinkers smarter than the best human geniuses, art generators better than the best human virtuosos, et cetera. Hell, the AIs at the centre of this bubble are running on [neural networks](https://en.wikipedia.org/wiki/Neural_network_(machine_learning)), whose functioning is based on our current understanding of What we *instead* got was Google telling us to [eat rocks and put glue in pizza](https://www.bbc.co.uk/news/articles/cd11gzejgz4o), chatbots hallucinating everything under the fucking sun, and art generators drowning the entire fucking internet in pure unfiltered slop, identifiable in the uniquely AI-like errors it makes. And all whilst burning through truly unholy amounts of power and receiving frankly embarrassing levels of hype in the process. (Quick sidenote: Even a local model running on some rando's GPU is a power-hog compared to what its trying to imitate - digging around online indicates your brain uses only [20 watts of power](https://hypertextbook.com/facts/2001/JacquelineLing.shtml) to do what it does.) With the parade of artificial stupidity the bubble's given us, I wouldn't fault anyone for coming to believe the brain isn't like a computer at all. 2) Inhuman Learning Additionally, AI bros have repeatedly and incessantly claimed that AIs are creative and that they learn like humans, usually in response to complaints about the Biblical amounts of art stolen for AI datasets. Said claims are, of course, flat-out bullshit - last I checked, human artists only need a few references to actually produce something good and original, whilst your average LLM will produce nothing but slop no matter how many terabytes upon terabytes of data you throw at its dataset. This all arguably falls under the "Artificial Idiocy" heading, but it felt necessary to point out - these things lack the creativity or learning capabilities of humans, and I wouldn't blame anyone for taking that to mean that brains are uniquely *unlike* computers. 3) Eau de Tech Asshole Given how much public resentment the AI bubble has built towards the tech industry (which [I covered in my previous post](https://awful.systems/post/2031653)), my gut instinct's telling me that the IP metaphor is also starting to be viewed in a harsher, more "tech asshole-ish" light - not just merely a reductive/incorrect view on human cognition, but as a sign you put tech over human lives, or don't see other people as human. Of course, AI providing a general parade of the absolute worst scumbaggery we know (with [Mira Murati being an anti-artist scumbag](https://nitter.poast.org/tsarnick/status/1803920566761722166) and [Sam Altman being a general creep](https://www.theverge.com/2024/5/20/24161253/scarlett-johansson-openai-altman-legal-action) as the biggest examples) is probably helping that fact, alongside all the active attempts by AI bros to mimic real artists ([exhibit A](https://twitter.com/anukaakash/status/1806854002640081345), [exhibit B](https://twitter.com/GenelJumalon/status/1810815644331278576)).

    30
    33
    buttondown.email

    Whilst going through MAIHT3K's backlog, I ended up running across a neat little article theorising on the possible aftermath which left me wondering precisely what the main "residue", so to speak, would be. **The TL;DR:** To cut a long story *far* too short, Alex, the writer, theorised the bubble would leave a "sticky residue" in the aftermath, "coating creative industries with a thick, sooty grime of an industry which grew expansively, without pausing to think about who would be caught in the blast radius" and killing or imperilling a lot of artists' jobs in the process - all whilst [producing metric assloads of emissions](https://pivot-to-ai.com/2024/07/23/data-centers-risk-missing-us-climate-goals-especially-with-ai/) and pushing humanity closer to the apocalypse. **My Thoughts** Personally, whilst I can see Alex's point, I think the main residue from this bubble is going to be large-scale resentment of the tech industry, for three main reasons: 1) AI Is Shafting Everyone Its not just artists who have been pissed off at AI fucking up their jobs, whether [freelance](https://finance.yahoo.com/news/ai-doesn-t-kill-jobs-010000898.html?fr=sycsrp_catchall) or [corporate](https://archive.is/GGNoC) - as [Upwork, of all places, has noted](https://www.upwork.com/research/ai-enhanced-work-models) in their research, pretty much anyone working right now is getting the shaft: - Nearly half (47%) of workers using AI say they have no idea how to achieve the productivity gains their employers expect - Over three in four (77%) say AI tools have decreased their productivity and added to their workload in at least one way - Seventy-one percent are burned out and nearly two-thirds (65%) report struggling with increasing employer demands - Women (74%) report feeling more burned out than do men (68%) - **1 in 3 employees** say they will likely **quit their jobs** in the next six months because they are burned out or overworked (emphasis mine) [Baldur Bjarnason put it better than me when commenting on these results](https://www.baldurbjarnason.com/2024/the-other-ai-shoe-dropping/): > It’s quite unusual for a study like this on a new office tool, roughly two years after that tool—ChatGPT—exploded into people’s workplaces, to return such a resoundingly negative sentiment. > > But it fits with the studies on the actual functionality of said tool: the incredibly common and hard to fix errors, the biases, the general low quality of the output, and the often stated expectation from management that it’s a magic fix for the organisational catastrophe that is the mass layoff fad. > > Marketing-funded research of the kind that Upwork does usually prevents these kind of results by finessing the questions. They simply do not directly ask questions that might have answers they don’t like. > > That they didn’t this time means they really, really did believe that “AI” is a magic productivity tool and weren’t prepared for even the possibility that it might be harmful. Speaking of the general low-quality output: 2) The AI Slop-Nami The Internet has been flooded with AI-generated garbage. Fucking FLOODED. Doesn't matter where you go - [Google](https://www.bbc.co.uk/news/articles/cd11gzejgz4o), [DeviantArt](https://slate.com/technology/2024/05/deviantart-what-happened-ai-decline-lawsuit-stability.html), [Amazon](https://www.404media.co/ai-generated-mushroom-foraging-books-amazon/), [Facebook](https://www.404media.co/email/1cdf7620-2e2f-4450-9cd9-e041f4f0c27f/), [Etsy](https://uk.pcmag.com/ai/153201/etsy-doubles-down-on-pro-ai-art-policies-despite-calls-for-ai-ban), [Instagram](https://eu.usatoday.com/story/opinion/voices/2024/07/24/instagram-ai-art-hurts-artists-bookstagram/74456870007/), [YouTube](https://www.bbc.co.uk/newsround/66796495), [Sports Illustrated](https://www.pbs.org/newshour/economy/sports-illustrated-found-publishing-ai-generated-stories-photos-and-authors), fucking 99% of the Internet is polluted with it. Unsurprisingly, this utter flood of unfiltered unmitigated *endless trash* has sent AI's public perception [straight down the fucking toilet](https://www.baldurbjarnason.com/2024/slop-framing-failure-as-success/), to the point of spawning [an entire counter-movement](https://archive.is/n9eBg) against the fucking thing. Whether it be [Glaze](https://glaze.cs.uchicago.edu/) and [Nightshade](https://nightshade.cs.uchicago.edu/) directly sabotaging datasets, "[Made with Human Intelligence](https://substack.com/@bethspencer/note/c-58614935)" and "[Not By AI](https://notbyai.fyi/)" badges proudly proclaiming human-done production or Cara [blowing up by offering a safe harbour from AI](https://techcrunch.com/2024/06/06/a-social-app-for-creatives-cara-grew-from-40k-to-650k-users-in-a-week-because-artists-are-fed-up-with-metas-ai-policies/), its clear there's a *lot* of people out there who want abso-fucking-lutely nothing to do with AI in any sense of the word as a result of this slop-nami. 3) The Monstrous Assholes In AI On top of this little slop-nami, those leading the charge of this bubble have been generally godawful human beings. Here's a quick highlight reel: - [Microsoft’s AI boss thinks it’s perfectly okay to steal content if it’s on the open web](https://www.theverge.com/2024/6/28/24188391/microsoft-ai-suleyman-social-contract-freeware) - it was already a safe bet anyone working in AI thought that considering they stole from [literally everyone](https://www.404media.co/listen-to-the-ai-generated-ripoff-songs-that-got-udio-and-suno-sued/), whether [large](https://spectrum.ieee.org/midjourney-copyright) or [small](https://www.404media.co/email/e3836b26-6914-4c1c-a102-bf9735adc3de/), but its nice for someone to spell it out like that. - [The death of robots.txt](https://www.theverge.com/24067997/robots-txt-ai-text-file-web-crawlers-spiders) - OpenAI [publicly ignored it](https://www.businessinsider.com/openai-anthropic-ai-ignore-rule-scraping-web-contect-robotstxt), Perplexity [lied about their user agent to deceive it](https://rknight.me/blog/perplexity-ai-is-lying-about-its-user-agent/), and Anthropic [spammed crawlers to sidestep it](https://www.404media.co/websites-are-blocking-the-wrong-ai-scrapers-because-ai-companies-keep-making-new-ones/). [Results were predictable](https://www.theregister.com/2024/07/22/ai_training_data_shrinks/). [Mostly](https://pivot-to-ai.com/2024/07/19/ai-models-being-blocked-from-fresh-data-except-the-trash/). - [Mira Murati saying "some creative jobs shouldn't exist"](https://nitter.poast.org/tsarnick/status/1803920566761722166) - Ed Zitron called this "a declaration of war against creative labor" [when talking to Business Insider](https://www.businessinsider.com/openai-cto-mira-murati-creative-jobs-eliminated-ai-2024-6), which sums this up better than I ever could. - [Sam Altman's Tangle with Her](https://nitter.poast.org/BobbyAllyn/status/1792679435701014908) - Misunderstanding the basic message? [Check](https://archive.is/wdrFD). Bullying a dragon (to [quote TV Tropes](https://tvtropes.org/pmwiki/pmwiki.php/Main/BullyingADragon))? [Check](https://www.hollywoodreporter.com/business/business-news/scarlett-johansson-disney-settle-black-widow-lawsuit-1235022598/). Embarassing your own company? [Check](https://www.npr.org/2024/05/20/1252495087/openai-pulls-ai-voice-that-was-compared-to-scarlett-johansson-in-the-movie-her). Being a creepy-ass motherfucker? [Fucking check](https://www.bloodinthemachine.com/p/why-is-sam-altman-so-obsessed-with). - "[Can Artificial Intelligence Speak for Incapacitated Patients at the End of Life?](https://archive.is/DcDFb#selection-2158.0-2158.1)" - I'll let Amy and David [speak for me on this](https://pivot-to-ai.com/2024/07/31/can-artificial-intelligence-speak-for-incapacitated-patients-at-the-end-of-life-no-and-what-the-hell-is-wrong-with-you/) because ***WHAT THE FUCK***. I'm definitely missing a lot, but I think this sampler gives you a good gist of the kind of soulless ghouls who have been forcing this entire fucking AI bubble upon us all. **Eau de Tech Asshole** There are many things I can't say for sure about the AI bubble - when it will burst, how long and harsh the next AI/tech winter will be, what new tech bubble will pop up in its place (if any), etcetera. One thing I feel I *can* say for sure, however, is that the AI bubble and its myriad harms will leave a *lasting* stigma on the tech industry once it finally bursts. Already, it seems AI has a pretty hefty stigma around it - as Baldur Bjaranason noted when talking about when discussing [AI's sentiment disconnect between tech and the public](https://www.baldurbjarnason.com/2024/sentiment-disconnect/): > To many, “AI” seems to have become a tech asshole signifier: the “tech asshole” is a person who works in tech, only cares about bullshit tech trends, and doesn’t care about the larger consequences of their work or their industry. Or, even worse, aspires to become a person who gets rich from working in a harmful industry. > > For example, my sister helps manage a book store as a day job. They hire a lot of teenagers as summer employees and at least those teens use “he’s a big fan of AI” as a red flag. (Obviously a book store is a biased sample. The ones that seek out a book store summer job are generally going to be good kids.) > > *I don’t think I’ve experienced a sentiment disconnect this massive in tech before, even during the dot-com bubble.* On another front, there's [the cultural reevaluation of the Luddites](https://archive.is/VKgW8) - once brushed off as naught but rejectors of progress, they are now coming to be viewed as folk heroes in a sense, fighting against misuse of technology to disempower and oppress, rather than technology as a whole. There's also the [rather recent SAG-AFTRA strike](https://www.theregister.com/2024/07/26/hollywood_video_game_strike/) which kicked off just under a year after [the previous one](https://en.wikipedia.org/wiki/2023_SAG-AFTRA_strike), and was started for similar reasons - to protect those working in the games industry from being shafted by AI like so many other people. With how the tech industry was responsible for creating this bubble at every stage - research, development, deployment, the whole nine yards - it is all but guaranteed they will shoulder the blame for all that its unleashed. Whatever happens after this bubble, I expect hefty scrutiny and distrust of the tech industry for a long, long time after this. To quote @datarama, "[the AI industry has made tech synonymous with “monstrous assholes” in a non-trivial chunk of public consciousness](https://awful.systems/comment/4128699)" - and that chunk is not going to forget any time soon.

    18
    1

    I've been hit by inspiration whilst dicking about on Discord - felt like making some off-the-cuff predictions on what will happen once the AI bubble bursts. (Mainly because I had a bee in my bonnet that was refusing to fuck off.) 1) A Full-Blown Tech Crash Its no secret the industry's put all their chips into AI - basically every public company's chasing it to inflate their stock prices, NVidia's making money hand-over-fist playing gold rush shovel seller, and every exec's been hyping it like its gonna change the course of humanity. Additionally, [going by Baldur Bjarnason](https://softwarecrisis.dev/letters/ai-is-a-hail-mary-pass/), tech's chief goal with this bubble is to prop up the notion of endless growth so it can continue reaping the benefits for just a bit longer. If and when the tech bubble pops, I expect a full-blown crash in the tech industry ([much like Ed Zitron's predicting](https://www.wheresyoured.at/peakai/)), with revenues and stock prices going through the floor and layoffs left and right. Additionally, I'm expecting those stock prices will likely take a while to recover, if ever, as tech likely comes to be viewed either as a stable, mature industry that's no longer experiencing nonstop growth or as an industry experiencing a full-blown [malaise era](https://en.wikipedia.org/wiki/Malaise_era), with valuations and stock prices getting savaged as Wall Street comes to see tech companies as high risk investments at best and money pits at worst. (Missed this incomplete sentence several times) **Chance:** Near-Guaranteed. I'm pretty much certain on this, and expect it to happen sometime this year. 2) A Decline in Tech/STEM Students/Graduates Extrapolating a bit from Prediction 1, I suspect we might see a lot less people going into tech/STEM degrees if tech crashes like I expect. The main thing which drew so many people to those degrees, at least from what I could see, was the notion that they'd make you a lotta money - if tech publicly crashes and burns like I expect, it'd blow a major hole in that notion. Even if it doesn't kill the notion entirely, I can see a fair number of students jumping ship at the sight of that notion being shaken. **Chance:** Low/Moderate. I've got no solid evidence this prediction's gonna come true, just a gut feeling. Epistemically speaking, I'm firing blind. 3) Tech/STEM's Public Image Changes - For The Worse The AI bubble's given us a pretty hefty amount of mockery-worthy shit - Mira Murati [shitting on the artists OpenAI screwed over](https://nitter.poast.org/tsarnick/status/1803920566761722166), Andrej Karpathy [shitting on every movie made pre-'95](https://nitter.poast.org/tsarnick/status/1752581828962365635), Sam Altman claiming [AI will soon solve all of physics](https://nitter.poast.org/tsarnick/status/1806071104148271434), Luma Labs [publicly embarassing themselves](https://nitter.poast.org/LumaLabsAI/status/1800921393321934915), ProperPrompter [recreating motion capture, But Worse^tm](https://nitter.poast.org/ProperPrompter/status/1809206091764613487), Mustafa Suleyman [treating everything on the 'Net as his to steal](https://www.theverge.com/2024/6/28/24188391/microsoft-ai-suleyman-social-contract-freeware), [et cetera](https://nitter.poast.org/miramurati/status/1804567253578662264), *[et cetera](https://nitter.poast.org/MichaelG_3D/status/1804376761398104135)*, **et *[fucking](https://www.youtube.com/watch?v=ddTV12hErTc)* [cetera](https://www.youtube.com/watch?v=TitZV6k8zfA)**. All the while, AI has been [flooding the Internet](https://www.neilsahota.com/ai-slop-the-unseen-flood-of-ai-generated-content/) with [unholy slop](https://archive.ph/xva9j), [ruining Google search](https://www.techradar.com/computing/search-engines/google-search-might-be-getting-worse-and-ai-threatens-to-ruin-it-entirely), [cooking the planet](https://pivot-to-ai.com/2024/07/23/data-centers-risk-missing-us-climate-goals-especially-with-ai/), [stealing everyone's work](https://spectrum.ieee.org/midjourney-copyright) ([sometimes literally](https://nitter.poast.org/jakezward/status/1728032639402037610)) in [broad daylight](https://www.404media.co/listen-to-the-ai-generated-ripoff-songs-that-got-udio-and-suno-sued/), [supercharging scams](https://www.abc.net.au/news/2023-04-12/artificial-intelligence-ai-scams-voice-cloning-phishing-chatgpt/102064086), [killing livelihoods](https://archive.ph/GGNoC), [exploiting the Global South](https://archive.ph/cysp9) and God-knows-what-the-fuck-else. **All** of this has been a near-direct consequence of the development of large language models and generative AI. Baldur Bjarnason has already mentioned AI being [treated as a major red flag by many](https://www.baldurbjarnason.com/2024/sentiment-disconnect/) - a "tech asshole" signifier to be more specific - and the massive disconnect in sentiment tech has from the rest of the public. I suspect that "tech asshole" stench is gonna spread much quicker than he thinks. **Chance:** Moderate/High. This one's also based on a gut feeling, but with the stuff I've witnessed, I'm feeling much more confident with this than Prediction 2. Arguably, if the [cultural rehabilitation of the Luddites](https://archive.ph/VKgW8) is any indication, it might already be happening without my knowledge. If you've got any other predictions, or want to put up some criticisms of mine, go ahead and comment.

    15
    8
    www.theatlantic.com

    Damn nice sneer from Charlie Warzel in this one, taking a direct shot at Silicon Valley and its AGI rhetoric. [Archive link](https://archive.ph/hhAJ2), to get past the paywall.

    25
    0

    (Gonna expand on [a comment I whipped out yesterday](https://awful.systems/comment/4047633) - feel free to read it for more context) ---- At this point, its already well known AI bros are crawling up everyone's ass and scraping whatever shit they can find - [robots.txt](https://www.businessinsider.com/openai-anthropic-ai-ignore-rule-scraping-web-contect-robotstxt), [honesty](https://rknight.me/blog/perplexity-ai-is-lying-about-its-user-agent/) and [basic decency](https://rknight.me/blog/as-if-perplexity-didnt-suck-enough-theyre-also-hotlinking-images/) be damned. The good news is that services have started popping up to actively cockblock AI bros' digital smash-and-grabs - Cloudflare made waves when they [began offering blocking services for their customers](https://blog.cloudflare.com/declaring-your-aindependence-block-ai-bots-scrapers-and-crawlers-with-a-single-click), but Spawning AI's recently [put out a beta for an auto-blocking service of their own called Kudurru](https://kudurru.ai/). (Sidenote: Pretty clever of them to call it [Kudurru](https://en.wikipedia.org/wiki/Kudurru).) I do feel like active anti-scraping measures could go somewhat further, though - the obvious route in my eyes would be to try to actively feed complete garbage to scrapers instead - whether by sticking a bunch of garbage on webpages to mislead scrapers or by trying to [prompt inject the shit out of the AIs themselves](https://www.bentasker.co.uk/posts/blog/security/perplexity-ai-gives-answers-that-cannot-be-trusted.html). The main advantage I can see is subtlety - it'll be obvious to AI corps if their scrapers are given a 403 Forbidden and told to fuck off, but the chance of them noticing that their scrapers are getting fed complete bullshit isn't that high - especially considering AI bros aren't the brightest bulbs in the shed. Arguably, AI art generators are already getting sabotaged this way to a strong extent - [Glaze](https://glaze.cs.uchicago.edu/) and [Nightshade](https://nightshade.cs.uchicago.edu/) aside, ChatGPT et al's slop-nami has provided a lot of opportunities for AI-generated garbage (text, music, art, etcetera) to get scraped and poison AI datasets in the process. How effective this will be against the "summarise this shit for me" chatbots which inspired this high-length shitpost I'm not 100% sure, but between [one proven case of prompt injection](https://www.bentasker.co.uk/posts/blog/security/perplexity-ai-gives-answers-that-cannot-be-trusted.html) and [AI's dogshit security record](https://pivot-to-ai.com/2024/07/12/llm-vendors-are-incredibly-bad-at-responding-to-security-issues/), I expect effectiveness will be pretty high.

    20
    10

    After reading through Baldur's latest piece on [how tech and the public view gen-AI](https://www.baldurbjarnason.com/2024/sentiment-disconnect/), I've had some loose thoughts about how this AI bubble's gonna play out. I don't have any particular structure to this, this is just a bunch of things I'm getting off my chest: 1) AI's Dogshit Reputation Past AI springs had the good fortune to have had no obvious negative externalities to sour the public's reputation (mainly because they weren't public facing, [going by David Gerard](https://awful.systems/comment/3788633)). This bubble, by comparison, has been pretty much entirely public facing, giving us, among other things: - A veritable slop-nami of garbage-looking art, interesting only when it comes off as [completely fucking insane](https://nitter.poast.org/FacebookAIslop) (say hi [Biblically-accurate gymnasts](https://nitter.poast.org/d_feldman/status/1806950107872469069)) - Copyright infringement and art theft on a Biblical scale, leading to basically major AI company getting sued out the ass (with [Suno and Udio](https://www.billboard.com/pro/major-label-lawsuit-ai-firms-suno-udio-copyright-infringement/) being the latest targets) - Colossal amounts of power consumption, and thus planet-cooking levels of CO2 emissions (for the latest example, Google [missed its climate targets as a *direct result* of AI](https://www.euronews.com/green/2024/07/03/climate-scientists-urge-responsible-use-of-ai-as-googles-emissions-soar-by-48-since-2019)) - High-profile public embarrassments left and right, with [Google's pizza-glue pisstake](https://www.bbc.co.uk/news/articles/cd11gzejgz4o) the most obvious coming to mind - Scammers making use of voice-cloning tech to [make their scams more convincing](https://www.abc.net.au/news/2023-04-12/artificial-intelligence-ai-scams-voice-cloning-phishing-chatgpt/102064086) (with a particularly notorious flavour [imitating a loved one under duress](https://www.newyorker.com/science/annals-of-artificial-intelligence/the-terrifying-ai-scam-that-uses-your-loved-ones-voice)) (thanks to @mountainriver for pointing this one out) - And probably a few more I'm missing All of these have done a lot of damage to AI's public image, to the point where [its absence is an explicit selling point](https://archive.is/n9eBg) - damage which I expect to last for at least a decade. When the next AI winter comes in, I'm expecting it to be particularly long and harsh - I fully believe a lot of would-be AI researchers have decided to go off and do something else, rather than risk causing or aggravating shit like this. (Missed this incomplete sentence on first draft) 2) The Copyright Shitshow Speaking of copyright, basically every AI company has worked under the assumption that copyright basically doesn't exist and [they can yoink whatever they want without issue](https://www.theverge.com/2024/6/28/24188391/microsoft-ai-suleyman-social-contract-freeware). With Gen-AI being Gen-AI, getting evidence of their theft isn't particularly hard - as they're straight-up incapable of creativity, they'll puke out replicas of its training data with the right prompt. Said training data has included, on the audio side, [songs held under copyright by major music studios](https://www.404media.co/listen-to-the-ai-generated-ripoff-songs-that-got-udio-and-suno-sued/), and, on the visual side, [movies and cartoons currently owned *by the fucking Mouse.*](https://spectrum.ieee.org/midjourney-copyright). Unsurprisingly, they're getting sued to kingdom come. If I were in their shoes, I'd probably try to convince the big firms my company's worth more alive than dead and strike some deals with them, a la [OpenAI with Newscorp](https://www.theregister.com/2024/05/23/openai_news_corp/). Given they seemingly believe they did nothing wrong (or at least [Suno and Udio do](https://www.billboard.com/pro/ai-music-companies-hire-law-firm-defend-label-lawsuits/)), I expect they'll try to fight the suits, get pummeled in court, and almost certainly go bankrupt. There's also the [AI-focused COPIED act](https://www.theverge.com/2024/7/11/24196769/copied-act-cantwell-blackburn-heinrich-ai-journalists-artists) which would explicitly ban these kinds of copyright-related shenanigans - between getting bipartisan support and support from a lot of major media companies, chances are good it'll pass. 3) Tech's Tainted Image I feel the tech industry as a whole is gonna see its image get further tainted by this, as well - the industry's image has already been falling apart for a while, but it feels like AI's sent that decline into high gear. When the cultural zeitgeist is doing a 180 [on the fucking Luddites](https://archive.is/VKgW8) and is [openly clamoring for AI-free shit](https://archive.is/n9eBg), whilst Apple produces [the tech industry's equivalent](https://www.theverge.com/2024/5/9/24153113/apple-ipad-ad-crushing-apology) to the ["face ad"](https://en.wikipedia.org/wiki/1993_Chrétien_attack_ad), its not hard to see why I feel that way. I don't really know how things are gonna play out because of this. Taking a shot in the dark, I suspect the "tech asshole" stench Baldur mentioned is gonna be spread to the rest of the industry thanks to the AI bubble, and its gonna turn a fair number of people away from working in the industry as a result.

    16
    2
    www.baldurbjarnason.com

    > I don’t think I’ve ever experienced before this big of a sentiment gap between tech – web tech especially – and the public sentiment I hear from the people I know and the media I experience. > > Most of the time I hear “AI” mentioned on Icelandic mainstream media or from people I know outside of tech, it’s being used as to describe something as a specific kind of bad. “It’s very AI-like” (“mjög gervigreindarlegt” in Icelandic) has become the talk radio short hand for uninventive, clichéd, and formulaic. babe wake up the butlerian jihad is coming

    36
    5
    www.baldurbjarnason.com

    >I stopped writing seriously about “AI” a few months ago because I felt that it was more important to promote the critical voices of those doing substantive research in the field. > > But also because anybody who hadn’t become a sceptic about LLMs and diffusion models by the end of 2023 was just flat out wilfully ignoring the facts. > > The public has for a while now switched to using “AI” as a negative – using the term “artificial” much as you do with “artificial flavouring” or “that smile’s artificial”. > > But it seems that the sentiment might be shifting, even among those predisposed to believe in “AI”, at least in part. Between this, and [the rise of "AI-free" as a marketing strategy](https://twitter.com/bcmerchant/status/1801289558107308206), the bursting of the AI bubble seems quite close. Another solid piece from Bjarnason.

    39
    4