• theluddite@lemmy.ml
    link
    fedilink
    English
    arrow-up
    2
    ·
    1 year ago

    This is an article about a tweet with a screenshot of an LLM prompt and response. This is rock fucking bottom content generation. Look I can do this too:

    ![ChatGPT screenshot. prompt: Is chatgpt spreading misinformation? ChatGPT said: ChatGPT, like other AI models, can occasionally spread misinformation due to limitations in training data and interpretive algorithms. The model learns from a vast dataset but doesn’t actually “know” facts in the way humans do; instead, it generates responses based on patterns it has seen. If there are inaccuracies in the training data, those can sometimes be reflected in the responses.

    Additionally, for topics that require up-to-date information, an AI without real-time access to new data might give outdated answers. OpenAI, the organization behind ChatGPT, has worked to reduce misinformation by employing rigorous training processes, but no system is perfect. Fact-checking and user awareness remain essential when using AI for information.](https://lemmy.ml/pictrs/image/9559a382-6421-4116-8d88-9d7c13b7f044.png)

    • brucethemoose@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      1 year ago

      To add to this:

      All LLMs absolutely have a sycophancy bias. It’s what the model is built to do. Even wildly unhinged local ones tend to ‘agree’ or hedge, generally speaking, if they have any instruction tuning.

      Base models can be better in this respect, as their only goal is ostensibly “complete this paragraph” like a naive improv actor, but even thats kinda diminished now because so much ChatGPT is leaking into training data. And users aren’t exposed to base models unless they are local LLM nerds.

      • mm_maybe@sh.itjust.works
        link
        fedilink
        English
        arrow-up
        0
        ·
        1 year ago

        One of the reasons I love StarCoder, even for non-coding tasks. Trained only on Github means no “instruction finetuning” bullshit ChatGPT-speak.

          • mm_maybe@sh.itjust.works
            link
            fedilink
            English
            arrow-up
            1
            ·
            1 year ago

            I really wish it were easier to fine-tune and run inference on GPT-J-6B as well… that was a gem of a base model for research purposes, and for a hot minute circa Dolly there were finally some signs it would become more feasible to run locally. But all the effort going into llama.cpp and GGUF kinda left GPT-J behind. GPT4All used to support it, I think, but last I checked the documentation had huge holes as to how exactly that’s done.

    • Mac@mander.xyz
      link
      fedilink
      English
      arrow-up
      0
      ·
      edit-2
      1 year ago

      God, i love LLMs. (sarcasm)

      They will say anything you tell them to and you can even lead them into saying shit without explicitly stating it.
      They are not to be trusted.

      • essteeyou@lemmy.world
        link
        fedilink
        English
        arrow-up
        1
        ·
        1 year ago

        I tried it with your username and instance host and it thought it was an email address. When I corrected it, it said:

        I couldn’t find any specific information linking the Lemmy account or instance host “Mac@mander.xyz” to the dissemination of misinformation. It’s possible that this account is associated with a private individual or organization not widely recognized in public records.

        • Mac@mander.xyz
          link
          fedilink
          English
          arrow-up
          1
          ·
          1 year ago

          Right, because i told it to say that and left out the context. You can’t trust LLMs already and you must absolutely assume someone is lying or being disingenuous when all you have is a screenshot.