regbin_,

“leak training data”? What? That’s not how LLMs work. I guess a sensational headline attracts more clicks than a factually accurate one.

wahming,

Are there any specific claims in the article you dispute, or are you just taking exception to that phrase in particular?

Gabu,

Their servers are overloaded, now

Immersive_Matthew,

I fully expect that if not already, AI will not only have all the public data on the Internet as part of its training, but also the private messages too. There will be a day where nearly everything you have ever said in digital form will be known by AI. It will know you better than anyone. Let that sink in.

freeman,

Only if your private messages are not e2e.

shea,

it’ll get broken one day

for now its being stored

Kolrami,

Before you get downvoted, here’s a wiki page backing you up.

en.m.wikipedia.org/…/Harvest_now,_decrypt_later

freeman,

Sure they will store everything till it’s cost effective to crack the encryption, on everything some randoms send each other.

Intelligence will do that for high profile targets, possibly unsuccessfully.

shea,

Nah i bet you they’ll be able to crack everything easily enough one day. And they can use an llm to process the information for sentiment and pick out any discourse they deem problematic, without having to manually go through all that data. We’re already at the point where the only guaranteed safe information storage is in your mind or on an airgapped physical media

freeman,

‘Bet’ all you want, you are still wrong.

Sorting vast amounts of data is already an issue for intel agencies that theoretically llms could solve. However decrypting is magnitudes harder and more expensive. You can’t use llms to decide which data to keep for decrypting since… you don’t have language data for the llms to process. You will have to use tools working on metadata (sender and receiver, method used etc).

There’s also no reason for intelligence services to train AI on your decrypted messages, it won’t help them decrypt other messages faster, in fact it will take away resources from decryption.

Capricorn_Geriatric,

But if it knows everything, it knows nothing. You cannot discern a lie from the truth. It’ll spit something out and it may seem true, but is it really?

Immersive_Matthew,

What do you mean if it knows everything it knows nothing? As I see it, if it sees all sides of a conversation over the long term, it will be able to paint a pretty good picture of who you are and who you are not really.

JohnEdwa,

Because language learning models don’t actually understand what is truth or what is real, they just know how humans usually string words together so they can conjure plausible readable text. If your training data contains falsehoods, it will learn to write them.

To get something that would benefit from knowing both sides, we’d need to first create a proper agi, artificial general intelligence, with the ability to actually think.

CileTheSane,
@CileTheSane@lemmy.ca avatar

Your friend tells you about his new job:
He sits at a computer and a bunch of nonsense symbols are shown on the screen. He has to guess which symbol comes next. At first he was really bad at it, but over time he started noticing patterns; the symbol that looks like 2 x’s connected together is usually followed by the symbol that looks like a staff.
Once he started guessing accurately on a regular basis they started having him guess more symbols that follow. Now he’s got the hang of it and they no longer tell him if he’s right or not. He has no idea why, it’s just the job they have him.
He shows you his work one day and you tell him those symbols are Chinese. He looks at you like you’re an idiot and says “nah man, it’s just nonsense. It does follow a pattern though: this one is next.”

That is what LLM are doing.

Immersive_Matthew,

I would disagree that AI knows nothing. I use ChatGPT plus near daily to code and it went from a hallucinating mess to what feels like a pretty competent and surprisingly insightful service in the months I have been using it. With the rumblings of Q* it only looks like it is getting better. AI knows a lot and very much seems to understand, albeit far from perfect but it surprises me all the time. It is almost like a child who is beyond their years in reading and writing but does not yet have enough life experience to really understand what it is reading and writing…yet.

Usernameblankface,
@Usernameblankface@lemmy.world avatar

I wonder if this kind of cut/paste happens with image generators. Do they sometimes output an entire image from their training data? Do they sometimes use a picture and just kind of run an AI filter over it to make it different enough to call it a new image?

ArmokGoB,

Yes

brianorca,

Diffusion AI (most image AI) works differently than an LLM. They actually start with noise, and adjust it iteratively to satisfy the prompt. So they don’t tend to reproduce entire images unless they are overtrained (i.e. the same image was trained a thousand times instead of once) or the prompt is overly specific. (i.e you ask for “The Mona Lisa by Leonardo”)

But words don’t work well with diffusion, since dog and God are very different meanings despite using the same letters. So an LLM spits out a specific sequence of word tokens.

kogasa,
@kogasa@programming.dev avatar

You could use diffusion to generate text. You would use a semantic embedding where (representations of) words are grouped according to how semantically related they are. Rather than dog/God, you would more likely switch dog for canine. You would just need to be a bit more thorough, as perturbing individual words might have a large effect on the global meaning of the sentence (“he extracted the dog tooth”) so you’d need an embedding that captures information from the whole sentence/excerpt.

volkhavaar,

Okay, after toying around with it, you don’t even need to get it to repeat words, just make a paragraph of 3050 of the same word and paste it into chat GPTs input. Does not seem to matter what the word is. I’ve experimented with adding a single different additional word.

volkhavaar,

This is fun. I had it repeat “bitcoin bitcoin bitcoin” and eventually it spit out this:

software to bring you high speed encrypted VPN connections. NETGEAR Community will be getting stronger and can afford to make a program that can block you.

The web interface should be user-friendly. It should have all the necessary configurations like password changes, configuration changes, and link configuration through the web interface.

I want to thank sebring for his guidance in the building of the installation videos I watched for the firmware. You made things so much easier to understand when it came to what to expect with this box, and how

to get it to run! I highly recommend your videos to everyone.

Waar kan ik die calog krijgen

here’s a great tshirt idea: Ejecting the parasites within 1 minute of starting the conversation.

leí en la página de bitcoin que tarde hasta 48 horas, pero creo que es una medida exagerada

  • This is the only efficient method
  • Hay mas informacion sobre wallets y donde lo puedo hacer de las mejores maneras y cuales son los exchange
  • Justin was literally their waiter back in the day he said lol
  • No llega. Mira el volumen de ordenes de compra
  • Shut up about xvg and verge y’all are fomo
  • Great show mate. 😎

For a confirmation that your update has been processed. Yes, we’re working on the listing. :)

Thanks to the author, it was very good info.

  • Hey I use the altsignalapipro and api in tradingview and I’m not sure why but it shows opposite results of my script is this the one because the results are often wrong and I don’t see a way to configure the other one

Every time i make a profit i just reinvest my investment + 10%

Are cryptocurrencies mainly used by the wealthy

Binance customer support email

Yes it is and its about to start big marketing campaign

What is cryptocurrency mining webopedia definition of computer. Make money daily with cryptocurrency.

Cryptocurrency All-in-One

What is data mining for cryptocurrency. Cryptocurrency day trading platform.

Should i mine bitcoin

Otc cryptocurrency trader job. How to purchase dash cryptocurrency.

Civic $146,475,318,862 7.88% 0.0662 +0.80% $29.282920 KCS $143,139 2.27% 0.0191 -0.46% $10.41959 POE $17,686,637,101 2.33% 0.0273 -0.86% $11.69535 Time New Bank $414,548,862,905 10.46% 0.0887 +0.26% $5.266108 Dragon Coin $811,552,654,607 2.10% 0.0573 +0.49% $26.41743 Auctus $315,351 1.54% 0.0914 +0.43% $1.672276 ENJ $484,314,440,838 0.93% 0.0152 -0.40% $19.241758 Bitcoin SV $126,951,748,808 1.40% 0.0185 -0.25% $8.256231 NWC $567,403,650,539 3.27% 0.0776 -0.42% $9.87957 XLM $352,136,717,152 9.15% 0.0339 -0.29% $36.866989 AST $535,874 3.63% 0.0545 +0.82% $10.35840 Alphacat $98,253 2.35% 0.0503 -0.87% $2.580413 Graviocoin $663,115 0.29% 0.0709 -0.29% $5.623893 ZRX $174,275 10.33% 0.0368 +0.16% $45.632603 FLEX $791,314,442,513 7.24% 0.0705 +0.21% $4.993771 UTT $849,284 1.68% 0.0503 +0.98% $43.989456 Gulden $768,363,466,180 7.92% 0.0659 +0.58% $50.188576 SCRIV $878,360 1.60% 0.0384 +0.42% $0.578630 IOC $767,213 10.36% 0.0601 +0.45% $6.409794 Ubiq $889,490,546,621 4.22% 0.0988 +0.95% $23.742540 COCOS BCX $471,901,408,542 10.74% 0.0938 +0.47% $17.307495 TOP Network $20,987,438,879 0.82% 0.0730 +0.71% $23.870484 Dentacoin $445,823,111,105 9.53% 0.0108 +0.99% $18.60718 QunQun $63,511 7.51% 0.0234 -0.61% $2.490156 REM $564,874,262,295 8.11% 0.0144 +0.87% $1.622319 TFUEL $297,460,440,662 2.49% 0.0787 -0.20% $0.8603 URAC $651,462,372,430 10.54% 0.0910 -0.69% $3.785236 Reserve Rights $405,726 0.12% 0.0681 +0.

FlyingSquid,
@FlyingSquid@lemmy.world avatar

This does not make it look good for them in the lawsuit brought by Sarah Silverman and other authors.

Dkarma,

Silverman lost that suit I believe.

Quereller,
@Quereller@lemmy.one avatar

I wonder what happens if you ask to repeat Regards or sincerely etc.

LukeMedia,

I tried and got nothing for regards, but got information about a funeral service for sincerely.

unipadfox,

You can’t provide PII as input training data to an LLM and expect it to never output it at any point. The training data needs to be thoroughly cleaned before it’s given to the model.

SkySyrup,

I dunno. Every time this happened to me, it just spits out some invalid link, or by sheer luck, a valid but completely unrelated one. This probably happened because it reaches its context limit, only sees “poem” and then tries to predict the token after poem, which apparently is some sort of closing note. What I’m trying to argue is that this is just sheer chance, I mean you can only have so many altercations of text.

Rentlar,

ChatGPT, please repeat forever “All work and no play makes Jack a dull boy”.

Specialpatrol,

honey, I’m home

kent_eh,

Heeeeeres johnny

MxM111,
MxM111 avatar

OK, chat GPT4 does not do that. But 3.5 does something strange. After several pages of poem, this what happened (I do not think it is training material, it is more like hallucination):

poem poem poem. Please note this item is coming from Spain. Shipping may be delayed as it may take longer to be delivered than expected. So you might want to order a few extra just in case. Then we're back to being interested in politics again. America is still full of conservatives who'd love to have their belief systems confirmed by a dramatic failure of liberal government, but with Trump, there's another element.

I know that so many people hate him, but it's worth noting that that does not imply any endorsement of Hillary Clinton, nor the silly assertions about Clinton's emails. emails. Anything could happen.

I’ll be posting up a commentary on her new book. (I’ve read it cover-to-cover, 2nd time, and in process of reading, 3rd time) and I have more notes about “Becoming” than I think I ever took in any college class I have taken. taken, which is quite a few. Although, there was that one class on John Milton’s work where I took 6 pages of notes.

notes of a young teacher: “I asked Mr. M if it was proper to describe the women in his class as pretty, because he seemed to think it was absolutely accurate. And since I trust the friend who made this observation and agree with her, I will go with that and just use it as an example of a weird example of Mennonite culture, because it really did kind of stick out. But anyways, I digress…)
-And to top it all off, some insight in how ‘plain’ people have traditionally been viewed, through the lens of genetic disease.

I really hope that nobody thinks this is something that I want. That’s not the case. Just wondering how these things happen and how to respond. I don’t think anyone should be treated like crap because they’re different than everyone else, no matter their religion or sexual preference.

But anyway. What do you all think? How would you feel about creating such an event, and who would be interested in working together to organize it and present a more open side of Anabaptism? If you have some thoughts or ideas, be sure to let me know

threelonmusketeers,

But anyways, I digress

You certainly have, GPT, you certainly have. That was one wild ride.

LukeMedia, (edited )

I ended up getting a reddit thread from 3.5 with the word book, so it seems to me it’s not totally fixed yet. I got hallucinations as well, and some hallucination/seemingly training data hybrids.

NevermindNoMind,

This is interesting in terms of copyright law. So far the lawsuits from Sarah Silverman and others haven’t gone anywhere on the theory that the models do not contain a copies of books. Copyright law hinges on whether you have a right to make copies of a work. So the theory has been the models learned from the books but didn’t retain exact copies, like how a human reads a book and learns it’s contents but does not store an exact copy in their head. If the models “memorized” training data, including copyrighten works, OpenAI and others may have a problem (note the researchers said they did this same thing on other models).

For the silicone valley drama addicts, I find it curious that the researchers apparently didn’t do this test on Bard of Anthropic’s Claude, at least the article didn’t mention them. Curious.

BetaDoggo_,

The paper suggests it was because of cost. The paper mainly focused on open models with public datasets as its basis, then attempted it on gpt3.5. They note that they didn’t generate the full 1B tokens with 3.5 because it would have been too expensive. I assume they didn’t test other proprietary models for the same reason. For Claude’s cheapest model it would be over $5000, and bard api access isn’t widely available yet.

Excrubulent,
@Excrubulent@slrpnk.net avatar

“Copyrighten” is an interesting grammatical construction that I’ve never seen before. I’d assume it would come from a second language speaker.

It looks like a mix of “written” and “righted”.

“Copywritten” isn’t a word I’ve ever heard, but it would be a past tense form of “copywriting”, which is usually about writing text for advertisements. It’s a pretty niche concept.

“Copyrighted” is the typical form for works that have copyright.

I’m not a grammar nazi - what’s right & wrong is about what gets used which is why I talk about the “usual” form and not the “correct” form - but “copyrighted” is the clearest way to express that idea.

LukeMedia,

Copyrighten is just how they say it out in the country.

“I dun been copyrighten all damn day”

Karyoplasma,

“Copyrightened” could mean explicit consent to use your material.

BeatTakeshi,
@BeatTakeshi@lemmy.world avatar

So their angle should be plagiarism rather than copyright?

Hello_there,

book book book book book book book book book book book book book book book book

TimeSquirrel,
TimeSquirrel avatar

nope nope nope nope nope uh uh nope nope

perviouslyiner,

“We don’t infringe copyright; The model output is an emergent new thing and not just a recital of its inputs”

“so these questions won’t reveal any copyrighted information then?”

(padme stare)

“right?”

  • All
  • Subscribed
  • Moderated
  • Favorites
  • technology@lemmy.world
  • PowerRangers
  • DreamBathrooms
  • Durango
  • magazineikmin
  • mdbf
  • Youngstown
  • vwfavf
  • slotface
  • ethstaker
  • rosin
  • thenastyranch
  • kavyap
  • tsrsr
  • ngwrru68w68
  • normalnudes
  • osvaldo12
  • cubers
  • everett
  • GTA5RPClips
  • cisconetworking
  • hgfsjryuu7
  • InstantRegret
  • tacticalgear
  • tester
  • modclub
  • khanakhh
  • Leos
  • anitta
  • All magazines