• MrWiggles@prime8s.xyz
    link
    fedilink
    English
    arrow-up
    6
    ·
    edit-2
    1 year ago

    In the end, it’s important for those who make lying machines to be held accountable for the falsehoods they spread. Not to mention the copyright infringement. If I write code and it uses it for training, if it’s licensed under the GPL, all derivative code is copyleft. Am I wrong there? I know Microsoft trains Github Copilot on public code, does that mean they’re respecting the license of the code they’re using for training? I have so many questions.

  • AndrasKrigare@beehaw.org
    link
    fedilink
    English
    arrow-up
    4
    ·
    1 year ago

    I haven’t actually used ChatGPT, but I’ve used the “new bing” AI. With that, whenever it states something as fact, it fixes a citation to a website where it pulled it, so you can verify things. Does ChatGPT do that as well?

    • LittlePrimate@feddit.de
      link
      fedilink
      English
      arrow-up
      4
      ·
      1 year ago

      No. Chat-GPT is not sourcing it’s claims. I think it would also change it’s (usually misunderstood) purpose. Chat-GPT is a language model made to create responses that appear natural. It’s purpose was not to recite facts, although it often does so as a side effect on how it was trained, it simply creates likely word combinations. The researchers entered millions of texts and Chat-GPT ran some math to figure out which word is mostly likely to come next after each other word. So, simplified, it opperates on a likelyhood table of word relationships, generated when it was trained. This includes following up “Super Mario is” with “a video game character” as most texts it saw will refer to him as that. People mistake this as it generating facts (because if asked about things a factual response is likely because it’s what Chat-GPT usually saw) but this was never the purpose of Chat-GPT. So a response like “Super Mario is an orange cat who loves lasagne” would also be valid output, as it perfectly resembles natural language. It’s factually wrong but a correct sentence and after the first switch from “video game character” to cat, following up “orange cat” with love for lasagne is again a likely sentence. This is also what happens in most “made up” sentences. Chat-GPT takes a wrong turn somewhere (maybe because it does not have facts on a person or thing, maybe because it’s ambiguous or maybe because the user was actually trying to direct it into that sentence) but then continues to follow up with likely words. And once you realise that Chat-GPT always tries to create a natural, logical response, you can easily trick it into making up certain things. So if you ask about lawsuits regarding a certain person, it’ll create a natural sentence describing a lawsuit that this person was involved in that is entirely made up. But most importantly: the text will be grammatically correct and appear natural. And if Chat-GPT would respond “This person was never involved in a lawsuit” you can often simply say “OK, but let’s pretend this person was” and Chat-GPT will happily make something up.

      Bing’s purpose is different. It of course also has that “natural language” approach but additionally might actually run a web search to be able to quote and source it’s claims, whereas Chat-GPT does not even know what’s written on todays websites. It has access to it’s initial training data from September 2021 and a few additional datasets they used since then for refinement, but it has no live access to the Internet to look up up-to-date information. So Bing likely will be able to summarise the last Apple announcements where Chat-GPT will just say "sorry, I do not have that information ". If pushed, Chat-GPT might make up correct natural language sentences about that conference but the statement will be just be likely word combinations, not facts.

      • AndrasKrigare@beehaw.org
        link
        fedilink
        English
        arrow-up
        1
        ·
        1 year ago

        Interesting, I had assumed it was also doing some kind of live query when asked. Given that, I’m surprised there’s as much conversation and popularity as opposed to “new bing,” which I’m sure is a huge irritation for Microsoft. I know it has a crappy name, and “bing” doesn’t really have a sexy connotation anyways, but it seems like it provides a lot more of what people actually want from ChatGPT (although my understanding is new bing is based on ChatGPT v3 so in some ways is ChatGPT)

      • Pete Hahnloser@beehaw.orgOP
        link
        fedilink
        arrow-up
        3
        ·
        1 year ago

        That was my thinking. OpenAI may feel their disclaimers are airtight, but courts get the final say on such things. This sort of thing was bound to happen, and I’m glad no one suffered physical injury or death for the first lawsuit to be filed.

        • Pēteris Krišjānis@toot.lv
          link
          fedilink
          arrow-up
          3
          ·
          edit-2
          1 year ago

          @Powderhorn lot of tech billionaires live in wishful thinking world where they disrupt and void real world - including legal - consequences. They usually come crashing down. It is just surprising that big swaths of media absolutely never learn from this.

            • Pēteris Krišjānis@toot.lv
              link
              fedilink
              arrow-up
              1
              ·
              1 year ago

              @Powderhorn yeah, there is lot of sort of self-justification happening there in a nutshell.
              It is not that they are unique with this. Overall humanity still learns how to deal with their own quirks.

              • Pete Hahnloser@beehaw.orgOP
                link
                fedilink
                arrow-up
                1
                ·
                1 year ago

                Unrelated, is it considered proper etiquette to @ users one is replying to every time? I have no issue doing so; I just didn’t see a need. But it’s been pointed out that comment threads aren’t rendered the same for all users, so I can certainly start doing so.

                • Pēteris Krišjānis@toot.lv
                  link
                  fedilink
                  arrow-up
                  3
                  ·
                  1 year ago

                  @Powderhorn as far as I have seen, Fediverse somehow tracks thread without mentions. So for me it is up to someone’s preference :) All Mastodon clients I use prefix messages with mentions of persons I reply to, but they don’t have to.

  • ZappySnap@kbin.social
    link
    fedilink
    arrow-up
    1
    ·
    1 year ago

    I think any lawyer using a tool like this and not pulling the actual relevant case law is simply a moron. I’m not a lawyer, but I am a licensed engineer, and if I used ChatGPT to help kickstart any design thing, I’d take that result with an entire block of salt, and double check everything I was using, because ultimately I am responsible for the design. Same here with the lawyers.