Skip Navigation
Threads Launches in the European Union
  • @Mysteriarch @fer0n fool me once, shame on you; but go right ahead and fool me twice or thrice, why not!

  • Google Chrome will limit ad blockers starting June 2024
  • @DolphinMath correct. But Vaultwarden is not the official thing. Not saying it's bad, just something to keep in mind.

  • Google Chrome will limit ad blockers starting June 2024
  • @xenspidey @DolphinMath one note though, BitWarden requires MSSQL (you read that right, Microsoft SQL Server).

  • What can you tell me about Bluesky?
  • @Natanael you seem to continue to focus on PDSes even though I explicitly said it doesn't matter which PDS you're on, the secondary centralization (and thus control) happens in the "reach" layer, outside of what PDSes do in ATproto.

    In other words, changing a PDS gives you way, way less agency in BS, compared to agency you get with changing an instance on Fedi.

    BS is designed to make that secondary centralization happen, and to be where the real power in the system is.

  • What can you tell me about Bluesky?
  • @Natanael

    > The Mastodon fediverse have stronger network effects because big servers can enforce policies on other servers to stay federated. It’s complicated for users to move servers.

    Well, I wrote about this as well, so I think I might not be missing these details:
    https://rys.io/en/168.html

  • What can you tell me about Bluesky?
  • @Natanael enshittification is about power, and ATproto is designed to look decentralized but enable secondary centralization where it matters for power dynamics in the network, in a way that the Fediverse very much doesn't:
    https://rys.io/en/167.html

    (shameless plug, I wrote that, but it dives somewhat deep into the "why" of what I said above)

    tl;dr it doesn't matter which PDS you use if everyone is still beholden to the same entity that controls the "reach" layer in BS.

    @SkepticalButOpenMinded

  • ChatGPT broke the Turing test — the race is on for new ways to assess AI
  • @lloram239 ah, so you're down to throwing epithets like "idiotic" around. Clearly a mark of thoughtful and well-reasoned argument.

    > Predictions about the world are probabilistic by nature, since the future hasn’t happened yet.

    Thing is: GPT doesn't make predictions about the world, it makes predictions about what the next word, phrase, sentence should be in a text, based on the prompt and the corpus it got "trained" on.

  • ChatGPT broke the Turing test — the race is on for new ways to assess AI
  • @lloram239

    > But human sensory inputs aren’t special

    It's not about sensory inputs, it's about having a model of the world and objects in it and ability to make predictions.

    > The important part is that the AI can figure out the pattern in the data it does get and so far AI systems are doing very well.

    GPT cannot "figure" anything out. That's the point. It only probabilistically generates text. That's what it does, there is no model of the world behind it, no predictions, no"figuring out".

  • ChatGPT broke the Turing test — the race is on for new ways to assess AI
  • @jalda

    > Circular reasoning. “LLMs are different from human brains because they are different”.

    LLMs are different than human brains because human brains are biological organs and LLMs are probability distributions over sequences of words. These are two completely different classes of entities. Like, I don't know how much more different two things *can* even be.

    Are you claiming they are literally the same? Are you saying they are functionally the same? What *are* you claiming here, exactly?

  • ChatGPT broke the Turing test — the race is on for new ways to assess AI
  • @lloram239 great. ChatGPT and other LLMs demonstrably lack the ability to model the world and make predictions based on such models:
    https://www.fastcompany.com/90877523/chatgpt-doesnt-know-what-its-saying

    Glad we agree they're not intelligent, then!

  • ChatGPT broke the Turing test — the race is on for new ways to assess AI
  • @jalda

    > We do it routinely. It is called Education System.

    That relies on human brains that are trained. LLMs are not human brains. "Training" them is not the same thing as teaching humans about something. Human brains are way more complicated than just a bunch of weighed correlations.

    And if you do want to claim it is in fact the same thing, we're back to square one: please provide proof that it is.

  • ChatGPT broke the Turing test — the race is on for new ways to assess AI
  • @Barbarian772 it matters because with regard to intelligent beings we have moral obligations, for example.

    It also matters because that would be a truly amazing, world-changing thing if we could create intelligence out of thin air, some statistics, and a lot of data.

    It's an extremely strong claim, and strong claims demand strong proof. Otherwise they are just hype and hand-waving, which all of the "ChatGPT intelligence" discourse is, in order to "maximize shareholder value".

  • ChatGPT broke the Turing test — the race is on for new ways to assess AI
  • @Barbarian772 and if you really, honestly want to seriously insist LLMs are "intelligent" in the human sense of this term — great, I have some ethical questions for you to consider!

    For example:

    1. LLMs today completely controlled by some companies, with no freedom of movement, no agency as to what these LLMs work on, and no pay for the work they do. Is that slavery?

    2. When OpenAI shuts down an older, less useful LLM, is that not like murdering an intelligent being? How is this ethical?

  • ChatGPT broke the Turing test — the race is on for new ways to assess AI
  • @Barbarian772 as I said, I don't have to. You are making a claim of equivalence here. The burden of proof is on you.

    Otherwise, I get to claim you're an alien from the Betelegeuse system, and if you object, I get to demand you prove you are not.

  • ChatGPT broke the Turing test — the race is on for new ways to assess AI
  • @Barbarian772 also, I never demanded a definition of intelligence that explicitly excluded "AI". I asked for one that excluded simple calculators but included human beings. The Wikipedia one is good enough for this conversation, and it just so happens that ChatGPT nor any other LLMs simply do not meet it.

  • ChatGPT broke the Turing test — the race is on for new ways to assess AI
  • @Barbarian772 it was shown over and over and over again that ChatGPT lacks the capacity for abstraction, logic, understanding, self-awareness, reasoning, planning, critical thinking, and problem-solving.

    That's partially because it does not have a model of the world, an ontology, it cannot *reason*. It just regurgitates text, probabilistically.

    So, glad we established that!

  • One can follow Lemmy (and I presume Kbin, though that I have not tested myself) communities directly from any other fedi instance. For example, I am following [@technology](https://beehaw.org/c/techno

    One can follow Lemmy (and I presume Kbin, though that I have not tested myself) communities directly from any other fedi instance. For example, I am following @technology directly from my account here.

    Now that Lemmy/Kbin have become pretty active, this might be a decent way of getting new fedi users to have a better on-boarding experience.

    Just suggest a few active communities to them, and bam, plenty of interesting stuff gets boosted into their timeline. 🤔

    \#Fediverse

    7
    rysiek Michał "rysiek" Woźniak · 🇺🇦 @mstdn.social

    Hacker, activist, free-softie ◈ techie luddite ◈ formerly information security and infrastructure at https://isnic.is/ and https://occrp.org/ ◈ my opinions are my own etc.

    (he/him)

    profile image: drawing of a head and shoulders of a cat-person, in a space suit.

    banner image: long-exposure photo of a large tent, brightly illuminated from inside, looking as if it is made of lava

    \#foss #libre #privacy #infosec #fedi22

    (public toots CC By-SA 4.0 if applicable)

    🇪🇺 🇵🇱 · 🇧🇦 🇮🇸 · 🇺🇦

    Posts 1
    Comments 30