Natanael

@Natanael@slrpnk.net

Cryptography nerd

This profile is from a federated server and may be incomplete. Browse more on the original instance.

Natanael,

That’s not the take (although in a sense I agree training data should influence it especially if it materially reproduce training samples)

Instead the argument is that the individual outputs from ML can only be copyrighted if they carry a human expression (because that’s what the law is specifically meant to cover), if there’s creative height in the inputs to it resulting in an output carrying that expression.

Compare to photography - photographs aren’t protected automatically just because a button is pressed and an image is captured, rather you gain copyright protection as a result of your choice of motive which carries your expression.

Too simple prompts to ML models would under this ruling be considered to be comparable to uncopyrightable lists of facts (like a recipe) and thus the corresponding output is also not protected.

Natanael,

FYI there’s studies showing that repeatedly training ML on ML outputs degrades quality sharply after 4-5 cycles

Natanael,

It does after 4-5 cycles yeah, built in algorithmic biases gets amplified to a degree that it breaks the model

Natanael,

Both have similar purpose, although radically different implementations

Natanael,

Don’t forget that there can be multiple independent copyright claims. In the example of the painter, both the painter and the commissioner can own copyright to their respective contributions to the result.

Natanael,

TLDR write a creative poem as an input

Natanael,

You should do that on a server you control so you can set timestamps and all correctly while importing directly to the DB, not just automate posting of it all

Natanael,

Just need to post on Briar and scuttlebutt too and you’ve got them all, lol

(Funny seeing another bsky user in the wild, btw)

Natanael,

Not yet, they haven’t figured out how

Natanael,

That’s only true for constant rate compression, not for variable bitrate compression

Natanael,

What a weak tree, in my city the trees pretend the asphalt isn’t there and the roots grow right through it

Natanael,

The point is that ML can generate works in bulk outpacing humans by a ridiculous rate. And that it explicitly is meant to cover expression, but ML models don’t actually express anything, they just emit statistical averages of the input.

You don’t want a company with obviously way too few human employees to have created all of their works to be able to go look for similar art from others and threaten copyright lawsuits. By forcing humans to be involved in the process of creation you strongly limit the ability of such legal trolls to hurt other creators. Such copyright trolls ALREADY exists prior to ML, but extending copyright to unsupervised ML would superpower their lawsuits. They just have to spam various websites with some samples and then pretend everybody copied them.

This precedence is just what it should be. The reference to photos is completely correct. The creation of a sequence of bits isn’t in itself protected, it’s the selection of inputs in which the creativity lies that then carries over into protection of the output. Photos can be copyrighted because a human express something in their choice of motive. A surveillance camera for example don’t automatically give its operator copyright!

And it still doesn’t prevent you from using ML in the creation of things and claiming copyright, it just requires you to be the one directing the process instead of leaving it unsupervised.

Natanael,

Ok then, weighted averages.

Natanael,

Do I have to explain with math how my high level abstract reply applies?

Most generative ML rely on probabilities. The averages are over multidimensional complex data structures representing patterns extracted from the inputs. Like average faces when you prompt it for a face (try training it on different sets of faces and look at how the output differs, you really do see it retain averages of the patterns in the input such as average skin color and haircuts). I wasn’t talking about linear arithmetic averages.

Natanael, (edited )

You still misunderstand my use of “average”. I am once again not talking simple averages over simple arithmetic numbers.

Look at the outputs of models trained on majority white faces vs diverse faces. If you still don’t understand what I mean by averages then I guess this conversation is hopeless

Yes there’s noise in the process.

But that noise is applied in very specific ways, it still fundamentally tries to output what the training algorithm indicated you would expect from it given the prompt, staying in a general neighborhood preserving the syntax / structure / patterns in the input training data related to the keywords in your prompt. Ask for a face without more details and you get a face looking like the average face in the input, usually white in most models, western conventional haircuts, etc, because that’s representative of its inputs, an average over the extracted structure in the inputs. The noise just tweaks some selection of representative features and their exact properties. It is still close enough to average that I feel it is fair to call it average, because it so rarely output extremes (other than when the model just breaks down and produce nonsense).

Natanael,

You misunderstood again. The model isn’t creating the bias when it is trained on biased data. It just gives a representative output of its input. The average of many outputs will resemble the average of its inputs.

Natanael, (edited )

Recognition;

odsc.medium.com/the-impact-of-racial-bias-in-faci…

venturebeat.com/…/training-ai-algorithms-on-mostl…

Generative denoisers and colorization;

theverge.com/…/face-depixelizer-ai-machine-learni…

With generative models already used in stuff like adverts and soon in Hollywood, it becomes more relevant as it affects representation;

towardsdatascience.com/empowering-fairness-recogn…

This extends to text as the output more frequently copies a style which is common in the input.

Natanael,

Because the problem is explicitly about the choice between two doors. You have to eliminate all but two choices.

But even then, you’d still have a better chance by switching.

Your intuition about the change is the whole point - it exposes why the result is what it is.

Natanael,

You can do that more efficiently with transparency logs instead

Natanael,

But it’s achievable without any NFT

Natanael,

Meta, how does moderation / deletion work?

If i post to another instance and they remove my comment, does it also disappear from my own local profile view (as if the hosting instance controls display of all comments even remotely)? Or would it have to be removed by my own host to make it disappear from my own local profile view?

Natanael,

I was wondering where some of my comments went. Wouldn’t it be better if it would hide them from the thread view but let them stay in the profile (as long as your local mods doesn’t also remove)?

(And I don’t think those comments broke the rules and recieved no notification, so it seemed very weird that they just vanished)

Natanael,

At best I’m seeing a stub “load more comments” but then it doesn’t show either there OR in my own profile (where all other comments still show)

Natanael,

“no gatekeeping” as a response to telling them about literal genocide denial is the most absurd thing I’ve seen…

lemmy.blahaj.zone/modlog?page=1&userId=227710…

  • All
  • Subscribed
  • Moderated
  • Favorites
  • JUstTest
  • cubers
  • DreamBathrooms
  • InstantRegret
  • tacticalgear
  • magazineikmin
  • Youngstown
  • thenastyranch
  • mdbf
  • slotface
  • rosin
  • modclub
  • kavyap
  • ethstaker
  • megavids
  • osvaldo12
  • khanakhh
  • cisconetworking
  • Durango
  • everett
  • ngwrru68w68
  • Leos
  • normalnudes
  • GTA5RPClips
  • tester
  • anitta
  • provamag3
  • lostlight
  • All magazines