• 0 Posts
  • 20 Comments
Joined 2 years ago
cake
Cake day: July 29th, 2023

help-circle

  • It isn’t clear to me at this point that such research will ever be funded in english-speaking places without a significant set of regime changes… no politician or administrator can resist outsourcing their own thinking to llm vendors in exchange for funding. I expect the US educational system will eventually provide a terrible warning to everyone (except the UK, whose government looks at the US and says ā€œoh my god, that’s horrifying. How can we be more like that?ā€).

    I’m probably just feeling unreasonably pessimistic right now, though.



  • It is related, inasmuch as it’s all generated from the same prompt and the ā€œanswerā€ will be statistically likely to follow from the ā€œreasoningā€ text. But it is only likely to follow, which is why you can sometimes see a lot of unrelated or incorrect guff in ā€œreasoningā€ steps that’s misinterpreted as deliberate lying by ai doomers.

    I will confess that I don’t know what shapes the multiple ā€œlet me just checkā€ or correction steps you sometimes see. It might just be a response stream that is shaped like self-checking. It is also possible that the response stream is fed through a separate llm session when then pushes its own responses into the context window before the response is finished and sent back to the questioner, but that would boil down to ā€œneural networks pattern matching on each other’s outputs and generating plausible response token streamsā€ rather than any sort of meaningful introspection.

    I would expect the actual systems used by the likes of openai to be far more full of hacks and bodges and work-arounds and let’s-pretend prompts that either you or I could imagine.


  • It’s just more llm output, in the style of ā€œimagine you can reason about the question you’ve just been asked. Explain how you might have come about your answer.ā€ It has no resemblance to how a neural network functions, nor to the output filters the service providers use.

    It’s how the ai doomers get themselves into a flap over ā€œdeceptiveā€ models… ā€œomg it lied about its train of thought!ā€ because if course it didn’t lie, it just edited a stream of tokens that were statistically similar to something classified as reasoning during training.


  • I might be the only person here who thinks that the upcoming quantum bubble has the potential to deliver useful things (but boring useful things, and so harder to build hype on) but stuff like this particularly irritates me:

    https://quantumai.google/

    Quantum fucking ai? Motherfucker,

    • You don’t have ai, you have a chatbot
    • You don’t have a quantum computer, you have a tech demo for a single chip
    • Even if you had both of those things, you wouldn’t have ā€œquantum aiā€
    • if you have a very specialist and probably wallet-vaporisingly expensive quantum computer, why the hell would anyone want to glue an idiot chatbot to it, instead of putting it in the hands of competent experts who could actually do useful stuff with it?

    Best case scenario here is that this is how one department of Google get money out of the other bits of Google, because the internal bean counters cannot control their fiscal sphincters when someone says ā€œaiā€ to them.




  • I was reading a post by someone trying to make shell scripts with an llm, and at one point the system suggested making a directory called ~ (which is a shorthand for your home directory in a bunch of unix-alikes). When the user pointed out this was bad, the llm recommended remediation using rm -r ~ which would of course delete all your stuff.

    So, yeah, don’t let the approximately-correct machine do things by itself, when a single character substitution can destroy all your stuff.

    And JFC, being surprised that something called ā€œYOLOā€ might be bad? What were people expecting? --all-the-red-flags








  • For those of you who haven’t already seen it, r/accelerate is banning users who think they’ve talked to an AI god.

    https://www.404media.co/pro-ai-subreddit-bans-uptick-of-users-who-suffer-from-ai-delusions/

    There’s some optimism from the redditors that the LLM folk will patch the problem out (ā€œyou must be prompting it wrongā€), but assume that they somehow just don’t know about the issue yet.

    As soon as the companies realise this, red team it and patch the LLMs it should stop being a problem. But it’s clear that they’re not aware of the issue enough right now.

    There’s some dubious self-published analysis which coined the term ā€œneural howlroundā€ to mean some sort of undesirable recursive behaviour in LLMs that I haven’t read yet (and might not, because it sounds like cultspeak) and may not actually be relevant to the issue.

    It wraps up with a surprisingly sensible response from the subreddit staff.

    Our policy is to quietly ban those users and not engage with them, because we’re not qualified and it never goes well.

    AI boosters not claiming expertise in something, or offloading the task to an LLM? Good news, though surprising.




  • If it were merely a search engine, it risks not being ai enough. We already have search engines, and no one is gonna invest in that old garbage. So instead, it finds something that you might want that’s been predigested for ease of ai consumption (Retrieval), dumps it into the context window alongside your original question (Augmentation) and then bullshits about it (Generation).

    Think of it as exactly the same stuff that the LLM folk have already tried to sell you, trying to work around limitations of training and data availability by providing ā€œcut and paste as a serviceā€ to generate ever more complex prompts for you, in the hopes that this time you’ll pay more for it than it costs to run.


  • I like the idea of small communities, but a major issue (possibly the biggest issue) as demonstrated by many mastodon servers over the years is longevity. What happens when your admin gets bored/burns out/dies/goes fash/is replaced with an asshole/is unable or unwilling to moderate effectively?

    I don’t particularly like the big mastodon hosts (eg. mastodon.social) but they’re probably still going to be here tomorrow, unlike eg. octodon.social who are winding down because adminning was too much (after 8 years, which was a pretty good run!) and they didn’t have any plans or processes in place to handle this eventuality.

    Between that sort of thing and stuff like matrix cryptography being full of holes and large matrix room management being a nightmare and email really being gmail, I’m slowly coming round to the idea that federation is too hard to do well and that if we could just manage a decentralised identity service and decent client software then it wouldn’t matter if servers didn’t talk to each other because we’d still have 90% of what people wanted from federation in the first place. Just a simple matter of engineering, I’m sure.