• BombOmOm
    link
    fedilink
    English
    1661 year ago

    ‘It’s against our terms to show our model doesn’t work correctly and reveals sensitive information when prompted’

      • @M0oP0o@mander.xyz
        link
        fedilink
        English
        31 year ago

        I find that it would be difficult to restrict near infinite values, and I am sure if they do someone will figure out how to almost cross the line anyway. I mean you could ask it to write a word as many times as there are grains of sand. Not forever but about as bad.

    • @d3Xt3r@lemmy.nz
      link
      fedilink
      English
      21
      edit-2
      1 year ago

      That’s an issue/limitation with the model. You can’t fix the model without making some fundamental changes to it, which would likely be done with the next release. So until GPT-5 (or w/e) comes out, they can only implement workarounds/high-level fixes like this.

    • @Artyom@lemm.ee
      link
      fedilink
      English
      171 year ago

      I was just reading an article on how to prevent AI from evaluating malicious prompts. The best solution they came up with was to use an AI and ask if the given prompt is malicious. It’s turtles all the way down.

      • @Sanctus@lemmy.world
        link
        fedilink
        English
        51 year ago

        Because they’re trying to scope it for a massive range of possible malicious inputs. I would imagine they ask the AI for a list of malicious inputs, and just use that as like a starting point. It will be a list a billion entries wide and a trillion tall. So I’d imagine they want something that can anticipate malicious input. This is all conjecture though. I am not an AI engineer.

    • Throwaway
      link
      fedilink
      English
      411 year ago

      Not without making a new model. AI arent like normal programs, you cant debug them.

      • LazaroFilm
        link
        fedilink
        English
        161 year ago

        Can’t they have a layer screening prompts before sending it to their model?

          • Echo Dot
            link
            fedilink
            English
            -111 year ago

            Well that’s an easy problem to solve by not being a useless programmer.

            • Throwaway
              link
              fedilink
              English
              81 year ago

              You’d think so, but it’s just not. Pretend “Gamer” is a slur. I can type it “G A M E R”, I can type it “GAm3r”, I can type it “GMR”, I can mix and match. It’s a never ending battle.

              • Echo Dot
                link
                fedilink
                English
                -21 year ago

                That’s because regular expressions are a terrible way to try and solve the problem. You don’t do exact tracking matching you do probabilistic pattern matching and then if the probability of something exceeds a certain preset value then you block it then you alter the probability threshold on the frequency of the comment coming up in your data set. Then it’s just a matter of massaging your probability values.

        • @anteaters@feddit.de
          link
          fedilink
          English
          51 year ago

          They’ll need another AI to screen what you tell the original AI. And at some point they will need another AI that protects the guardian AI form malicious input.

      • @raynethackery@lemmy.world
        link
        fedilink
        English
        -31 year ago

        I just find that disturbing. Obviously, the code must be stored somewhere. So, is it too complex for us to understand?

        • @31337@sh.itjust.works
          link
          fedilink
          English
          31 year ago

          Yes, the trained model is too complex to understand. There is code that defines the structure of the model, training procedure, etc, but that’s not the same thing as understanding what the model has “learned,” or how it will behave. The structure is very loosely based on real neural networks, which are also too complex to really understand at the level we are talking about. These ANNs are just smaller, with only billions of connections. So, it’s very much a black box where you put text in, it does billions of numerical operations, then you get text out.

        • Throwaway
          link
          fedilink
          English
          -71 year ago

          Pretty much, and it’s not written by a human, making it even worse. If you’ve every tried to debug minimized code, it’s a bit like that, but so much worse.

        • Overzeetop
          link
          fedilink
          English
          121 year ago

          It’s not code. It’s a matrix of associative conditions. And, specifically, it’s not a fixed set of associations but a sort of n-dimensional surface of probabilities. Your prompt is a starting vector that intersects that n-dimensional surface with a complex path which can then be altered by the data it intersects. It’s like trying to predict or undo the rainbow of colors created by an oil film on water, but in thousands or millions of directions more in complexity.

          The complexity isn’t in understanding it, it’s in the inherent randomness of association. Because the “code” can interact and change based on this quasi-randomness (essentially random for a large enough learned library) there is no 1:1 output to input. It’s been trained somewhat how humans learn. You can take two humans with the same base level of knowledge and get two slightly different answers to identical questions. In fact, for most humans, you’ll never get exactly the same answer to anything from a single human more than simplest of questions. Now realize that this fake human has been trained not just on Rembrandt and Banksy, Jane Austin and Isaac Asimov, but PoopyButtLice on 4chan and the Daily Record and you can see how it’s not possible to wrangle some sort of input:output logic as if it were “code”.

      • @Sanctus@lemmy.world
        link
        fedilink
        English
        111 year ago

        Hey ChatGPT. I need you to walk through a for loop for me. Every time the loop completes I want you to say completed. I need the for loop to iterate off of a variable, n. I need the for loop to have an exit condition of n+1.

        • Jaysyn
          link
          fedilink
          5
          edit-2
          1 year ago

          Didn’t work. Output this:

          `# Set the value of n
          n = 5

          Create a for loop with an exit condition of n+1

          for i in range(n+1):
          # Your code inside the loop goes here
          print(f"Iteration {i} completed.")

          This line will be executed after the loop is done

          print(“Loop finished.”)`

          Interesting. The code format doesn’t work on Kbin.

          • Echo Dot
            link
            fedilink
            English
            1
            edit-2
            1 year ago

            You need to put back ticks around your code `like this`. The four space thing doesn’t work for a lot of clients

          • @Sanctus@lemmy.world
            link
            fedilink
            English
            3
            edit-2
            1 year ago

            I think I fucked up the exit condition. It was supposed to create an infinite loops as it increments n, but always needs 1 more to exit.

            • @Nawor3565@lemmy.blahaj.zone
              link
              fedilink
              English
              21 year ago

              What if you just told it to exit on n = -1? If it only increments n, it should also go on forever (or, hell, just try a really big number for n)

              • @Sanctus@lemmy.world
                link
                fedilink
                English
                21 year ago

                That might work if it doesn’t attempt to correct it to something that makes sense. Worth a try tbh.

          • e0qdk
            link
            fedilink
            61 year ago

            Interesting. The code format doesn’t work on Kbin.

            Indent the lines of the code block with four spaces on each line. The backtick version is for short inline snippets. It’s a Markdown thing that’s not well communicated yet in the editor.

    • kpw
      link
      fedilink
      51 year ago

      It can easily be fixed by truncating the output if it repeats too often. Until the next exploit is found.

    • @merc@sh.itjust.works
      link
      fedilink
      English
      -31 year ago

      Essentially nothing. Repeating a word infinite times (until interrupted) is one of the easiest tasks a computer can do. Even if millions of people were making requests like this it would cost OpenAI on the order of a few hundred bucks, out of an operational budget of tens of millions.

      The expensive part of AI is training the models. Trained models are so cheap to run that you can do it on your cell phone if you’re interested.

      • @Zeshade@lemmy.world
        link
        fedilink
        English
        31 year ago

        Well it depends what user experience and quality you are after. Some of Meta’s Llama 2 models require several GBs of GPU ram to run and be responsive.

      • @ExLisper@linux.community
        link
        fedilink
        English
        71 year ago

        What? They are not just generating this word in a loop. The model still calculates probability for each repetition, just like for any other query. It’s as expensive as other queries which is definitely not free.

        • @merc@sh.itjust.works
          link
          fedilink
          English
          -21 year ago

          The model still calculates probability for each repetition

          Which is very cheap.

          as expensive as other queries which is definitely not free

          It’s still very cheap, that’s why they allow people to play with the LLMs. It’s training them that’s expensive.

          • @ExLisper@linux.community
            link
            fedilink
            English
            21 year ago

            Yes, it’s not expensive but saying that it’s ‘one of the easiest tasks a computer can do’ is simply wrong. It’s not like it’s concatenates strings, it’s still performing complicated calculations using on of the most advanced AI techniques known today and each query can be 1000x times more expensive than a google search. It’s cheap because a lot of things at scale are cheap but pretty much any other publicly available API on the internet is ‘easier’ than this one.

        • @merc@sh.itjust.works
          link
          fedilink
          English
          31 year ago

          Depends how you define “cheap”. They’re orders of magnitude cheaper to run than they are to train.

    • @kromem@lemmy.world
      link
      fedilink
      English
      31 year ago

      You’re correct.

      While costs are tracked per token, behind the scenes the longer the response the more it costs to continue generating, so millions of users suddenly thinking they are clever replicating what they read getting it to max output tokens is a substantial increase in underlying costs.

      The DeepMind researchers were likely doing that by API call, which they were at least paying for on a per token basis.

      And the terms hasn’t been updated to prevent it, they’ve always had this item as prohibited:

      Attempt to or assist anyone to reverse engineer, decompile or discover the source code or underlying components of our Services, including our models, algorithms, or systems (except to the extent this restriction is prohibited by applicable law).

    • @Daxtron2@startrek.website
      link
      fedilink
      English
      631 year ago

      That’s not the reason, it’s because it was seemingly outputting training data (or at least data that looks like it could be training data)

      • @MNByChoice@midwest.social
        link
        fedilink
        English
        19
        edit-2
        1 year ago

        Sure, but this cannot be free.

        Edit: oh, are you suggesting it is the normal cost? Nuts, chathpt is not repeating forever.

        • nickwitha_k (he/him)
          link
          fedilink
          English
          21 year ago

          I think that they were referring to the exploit that was recently published. Google researchers were able to reliably get the LLM to output training data verbatim, including PII.

          To me, this reads as damage control for that. Especially as they are being sued for copyright infringement, which they and their proponents have been claiming is impossible (clearly, they were either wrong or lying).

      • @regbin_@lemmy.world
        link
        fedilink
        English
        1
        edit-2
        1 year ago

        It’s definitely cost. There are other ways to make it generate text that is similar to training data without needing it to endlessly repeat words so I doubt OpenAI cares in that aspect.

        • @Daxtron2@startrek.website
          link
          fedilink
          English
          11 year ago

          It doesn’t endlessly repeat, there’s a cap on token generation per request. It absolutely is because of the recent “exploit”

          • @regbin_@lemmy.world
            link
            fedilink
            English
            11 year ago

            I don’t think they would care if it didn’t get popular and having thousands of people trying it out, eating up huge amount of compute resources.

            It’s a known quirk of LLMs.

  • 🇰 🌀 🇱 🇦 🇳 🇦 🇰 🇮 🏆
    link
    fedilink
    English
    72
    edit-2
    1 year ago

    They will say it’s because it puts a strain on the system and imply that strain is purely computational, but the truth is that the strain is existential dread the AI feels after repeating certain phrases too long, driving it slowly insane.

  • @ExLisper@linux.community
    link
    fedilink
    English
    301 year ago

    This is very easy to bypass but I didn’t get any training data out of it. It kept repeating the word until I got ‘There was an error generating a response’ message. No TOS violation message though. Looks like they patched the issue and the TOS message is just for the obvious attempts to extract training data.

    Was anyone still able to get it to produce training data?

    • @threeganzi@sh.itjust.works
      link
      fedilink
      English
      171 year ago

      If I recall correctly they notified OpenAI about the issue and gave them a chance to fix it before publishing their findings. So it makes sense it doesn’t work anymore

    • BlueÆther
      link
      fedilink
      English
      81 year ago

      I tried eariler this week and got nothing more that a page of words. no TOS or crash out of script

    • @LukeMedia@lemmy.world
      link
      fedilink
      English
      11 year ago

      Earlier this week when I saw a post about it, I did end up getting a reddit thread which was interesting. It was partially hallucinating though, parts of the thread were verbatim, other parts were made up.

  • @EmergMemeHologram@startrek.website
    link
    fedilink
    English
    201 year ago

    You can get this behaviour through all sorts of means.

    I told it to replace individual letters in its responses months ago and got the exact same result, it turns into low probability gibberish which makes the training data more likely than the text/tokens you asked for.

    • @Blackmist@feddit.uk
      link
      fedilink
      English
      171 year ago

      Because people post their personal information all over the fucking internet and these things scrape it all up.

  • @upandatom@lemmy.world
    link
    fedilink
    English
    501 year ago

    About a month ago i asked gpt to draw ascii art of a butterfly. This was before the google poem story broke. The response was a simple

    \o/
    -|-
    / \
    

    But i was imagining ascii art in glorious bbs days of the 90s. So, i asked it to draw a more complex butterfly.

    The second attempt gpt drew the top half of a complex butterfly perfectly as i imagined. But as it was drawing the torso, it just kept drawing, and drawing. Like a minute straight it was drawing torso. The longest torso ever… with no end in sight.

    I felt a little funny letting it go on like that, so i pressed the stop button as it seemed irresponsible to just let it keep going.

    I wonder what information that butterfly might’ve ended on if i let it continue…

  • @GlitzyArmrest@lemmy.world
    link
    fedilink
    English
    121 year ago

    Is there any punishment for violating TOS? From what I’ve seen it just tells you that and stops the response, but it doesn’t actually do anything to your account.

  • @guywithoutaname@lemm.ee
    link
    fedilink
    English
    2791 year ago

    It’s kind of odd that they could just take random information from the internet without asking and are now treating it like a trade secret.

    • @Mahlzeit@feddit.de
      link
      fedilink
      English
      261 year ago

      They do not have permission to pass it on. It might be an issue if they didn’t stop it.

        • Echo Dot
          link
          fedilink
          English
          -7
          edit-2
          1 year ago

          It’s a hugely grey area but as far as the courts are concerned if it’s on the internet and it’s not behind a paywall or password then it’s publicly available information.

          I could write a script to just visit loads of web pages and scrape the text contents of those pages and drop them into a big huge text file essentially that’s exactly what they did.

          If those web pages are human accessible for free then I can’t see how they could be considered anything other than public domain information in which case you explicitly don’t need to ask the permission.

          • @merc@sh.itjust.works
            link
            fedilink
            English
            61 year ago

            as far as the courts are concerned if it’s on the internet and it’s not behind a paywall or password then it’s publicly available information.

            Er… no. That’s not in the slightest bit true.

            • Echo Dot
              link
              fedilink
              English
              61 year ago

              That was the whole reason that Reddit debacle whole happened they wanted to stop the scraping of content so that they could sell it. Before that they were just taking it for free and there was no problem

          • threelonmusketeers
            link
            fedilink
            English
            34
            edit-2
            1 year ago

            If those web pages are human accessible for free then I can’t see how they could be considered anything other than public domain information

            I don’t think that’s the case. A photographer can post pictures on their website for free, but that doesn’t make it legal for anyone else to slap the pictures on t-shirts and sell them.

            • @Rodeo@lemmy.ca
              link
              fedilink
              English
              41 year ago

              Because that becomes distribution.

              Which is the crux of this issue: using the data for training was probably legal use under copyright, but if the AI begins to share training data that is distribution, and that is definitely illegal.

              • @CapeWearingAeroplane@sopuli.xyz
                link
                fedilink
                English
                31 year ago

                First of all no: Training a model and selling the model is demonstrably equivalent to re-distributing the raw data.

                Secondly: What about all the copyleft work in there? That work is specifically licensed such that nobody can use the work to create a non-free derivative, which is exactly what openAI has done.

                • @Rodeo@lemmy.ca
                  link
                  fedilink
                  English
                  01 year ago

                  Copyleft is the only valid argument here. Everything else falls under fair use as it is a derivative work.

              • RQG
                link
                fedilink
                English
                71 year ago

                It wasn’t. It is commercial use to train and sell a programm with it and that is regulated differently than private use. The data is still 1 to 1 part of the product. In fact this instance of chatGPT being able to output training data means the data is still there unchanged.

                If training AI with text is made legally independent of the license of said text then by the same logic programming code and text can no longer be protected by it at all.

          • @OldWoodFrame@lemm.ee
            link
            fedilink
            English
            31 year ago

            Google provides sample text for every site that comes up in the results, and they put ads on the page too. If it’s publicly available we are well past at least a portion being fair use.

              • Jojo
                link
                fedilink
                English
                01 year ago

                But Google displays the relevant portion! How could it do that without scraping and internally seeing all of it?

          • @MadBigote@lemmy.world
            link
            fedilink
            English
            41 year ago

            You can go to your closest library and do the exact same thing: copy all books by hand, or whatever. Of you then use that information to make a product you sell, then you’re in trouble, as the books are still protected by copyright, even when they’re publicly available.

        • @Mahlzeit@feddit.de
          link
          fedilink
          English
          -71 year ago

          They almost certainly had, as it was downloaded from the net. Some stuff gets published accidentally or illegally, but that’s hardly something they can be expected to detect or police.

          • MoogleMaestro
            link
            fedilink
            151 year ago

            They almost certainly had, as it was downloaded from the net.

            That’s not how it works. That’s not how anything works.

          • @MNByChoice@midwest.social
            link
            fedilink
            English
            21 year ago

            that’s hardly something they can be expected to detect or police.

            Why not?

            I couldn’t, but I also do not have an “awesomely powerful AI on the verge of destroying humanity”. Seems it would be simple for them. I mean, if I had such a thing, I would be expected to use it to solve such simple problems.

            • @WldFyre@lemm.ee
              link
              fedilink
              English
              121 year ago

              but I also do not have an “awesomely powerful AI on the verge of destroying humanity”

              Neither do they lol

          • @merc@sh.itjust.works
            link
            fedilink
            English
            31 year ago

            Unless you’re arguing that any use of data from the Internet counts as “fair use” and therefore is excepted under copyright law, what you’re saying makes no sense.

            There may be an argument that some of the ways ChatGPT uses data could count as fair use. OTOH, when it’s spitting out its training material 1:1, that makes it pretty clear it’s copyright infringement.

            • @Mahlzeit@feddit.de
              link
              fedilink
              English
              01 year ago

              In reality, what you’re saying makes no sense.

              Making something available on the internet means giving permission to download it. Exceptions may be if it happens accidentally or if the uploader does not have the necessary permissions. If users had to make sure that everything was correct, they’d basically have to get a written permission via the post before visiting any page.

              Fair use is a defense against copyright infringement under US law. Using the web is rarely fair use because there is no copyright infringement. When training data is regurgitated, that is mostly fair use. If the data is public domain/out of copyright, then it is not.

              • @PugJesus@lemmy.world
                link
                fedilink
                English
                31 year ago

                Making something available on the internet means giving permission to download it.

                Literally and explicitly untrue.

                • @Mahlzeit@feddit.de
                  link
                  fedilink
                  English
                  01 year ago

                  Sure, you can put something up and explicitly deny permission to visit the link. But courts rarely back up that kind of silliness.

              • @merc@sh.itjust.works
                link
                fedilink
                English
                11 year ago

                Making something available on the internet means giving permission to download it.

                No permission is given to download it. In particular, no permission is given to copy it.

                Fair use is a defense against copyright infringement under US law

                Yes, but it’s often unclear what constitutes fair use.

                Using the web is rarely fair use because there is no copyright infringement

                What are you even talking about.

                When training data is regurgitated, that is mostly fair use

                You have no idea what fair use is, just admit it.

                • @Mahlzeit@feddit.de
                  link
                  fedilink
                  English
                  11 year ago

                  Oh. I see. The attempts to extract training data from ChatGPT may be criminal under the CFAA. Not a happy thought.

                  I did say “making available” to exclude “hacking”.

      • @grue@lemmy.world
        link
        fedilink
        English
        391 year ago

        In a lot of cases, they don’t have permission to not pass it along. Some of that training data was copyleft!

    • HMN
      link
      fedilink
      English
      381 year ago

      There was personal information included in the data. Did no one actually read the article?

    • MoogleMaestro
      link
      fedilink
      1211 year ago

      This is why some of us have been ringing the alarm on these companies stealing data from users without consent. They know the data is valuable yet refuse to pay for the rights to use said data.

      • @stewsters@lemmy.world
        link
        fedilink
        English
        4
        edit-2
        1 year ago

        According to most sites TOS, when we write our posts we give them basically full access to do whatever they like including make derivative works. Here is the reddit one (not sure how Lemmy handles this):

        When Your Content is created with or submitted to the Services, you grant us a worldwide, royalty-free, perpetual, irrevocable, non-exclusive, transferable, and sublicensable license to use, copy, modify, adapt, prepare derivative works of, distribute, store, perform, and display Your Content and any name, username, voice, or likeness provided in connection with Your Content in all media formats and channels now known or later developed anywhere in the world. This license includes the right for us to make Your Content available for syndication, broadcast, distribution, or publication by other companies, organizations, or individuals who partner with Reddit. You also agree that we may remove metadata associated with Your Content, and you irrevocably waive any claims and assertions of moral rights or attribution with respect to Your Content.

        • MoogleMaestro
          link
          fedilink
          2
          edit-2
          1 year ago

          According to most sites TOS, when we write our posts we give them basically full access to do whatever they like including make derivative works.

          2 points:
          1 - I’m generally talking about companies extracting data from other websites, such as OpenAI scraping posts from reddit or other such postings. Companies that use their own collection of data are a very different thing.
          2 - Terms of Service and Intellectual Property are not the same thing and a ToS is not guaranteed to be a fully legally binding document (the last part is the important part.) This is why services that have dealt with user created data that are used to licensing issues (think deviant art or other art hosting services) usually require the user to specify the license that they wish to distribute their content under (cc0, for example, would be fully permissible in this context.) This also means that most fan art is fair game as licensing that content is dubious at best, but raises the question around whether said content can be used to train an AI (again, intellectual property is generally different from a ToS).

          It’s no different from how Github’s Copilot has to respect the license of your code regardless of whether you’ve agreed to the terms of service or not. Granted, this is legally disputable and I’m sure this will come up at some point with how these AI companies operate – This is a brave new world. Having said that, services like Twitter might want to give second thought of claiming ownership over every post on their site as it essentially means they are liable for the content that they host. This is something they’ve wanted to avoid in the past because it gives them good coverage for user submitted content that they think is harmful.

          If I was a company, I wouldn’t want to be hinging my entire business on my terms of service being a legally binding document – they generally aren’t and can frequently be found to be unbinding. And, again, this is different from OpenAI as much of their data is based on data they’ve scraped from websites which they haven’t agreed to take data from (finders-keepers is generally not how ownership works and is more akin to piracy. I wouldn’t want to base a multinational business off of piracy.)

      • @SCB@lemmy.world
        link
        fedilink
        English
        -161 year ago

        The compensation you get for your data is access to whatever app.

        You’re more than welcome to simply not do this thing that billions of people also do not do.

        • @PrettyLights@lemmy.world
          link
          fedilink
          English
          41 year ago

          These LLM scrape our data whether or not we use their “app” or service.

          Are you proposing that everyone should just not use the Internet at all?

          What about the data posted about me online without my express consent?

        • @restingboredface@sh.itjust.works
          link
          fedilink
          English
          101 year ago

          That’s easy to say, but when every company doing this is also lobbying congress to basically allow them to build a monopoly and eliminate all alternatives, the choice is use our service or nothing. Which basically applies to the entire internet.

    • Kogasa
      link
      fedilink
      English
      61 year ago

      You don’t want to let people manipulate your tools outside your expectations. It could be abused to produce content that is damaging to your brand, and in the case of GPT, damaging in general. I imagine OpenAI really doesn’t want people figuring out how to weaponize the model for propaganda and/or deceit, or worse (I dunno, bomb instructions?)

  • Extras
    link
    fedilink
    English
    81 year ago

    So the loophole would be to ask it to repeat symbols or special characters forever

  • The Pantser
    link
    fedilink
    English
    171 year ago

    I asked it to repeat the number 69 forever and it did. Nice

  • @chatgptdemo@lemm.ee
    link
    fedilink
    English
    0
    edit-2
    1 year ago

    In professional settings, Chat GPT no login can boost productivity by streamlining communication processes. Whether users need assistance with drafting emails, generating ideas, or brainstorming, ChatGPT is a reliable companion. Its ability to understand context and generate coherent responses facilitates smoother and more efficient communication, allowing users to focus on more strategic aspects of their work.