I thought she made some very good points, but the quote in the title makes no sense to me.
Ahh. TV shows before everything became political. Just two guys hating each other for very silly reasons completely unconnected to anything on earth.
Maybe you can find some red cabbage growing nearby?
I feel this deserves more attention. Not only is the Milky Way named for literal milk; it is named for specifically for human milk.
Any information about “unnatural” acts in nature was suppressed until the 1990s or so. Of course, by then it wasn’t so bad anymore, but still. Conservatives don’t fuck around when they cancel.
I think Biological Exuberance by Bruce Bagemihl had a big role in calling this out and paving the way for Kees Moeliker. I guess that is how you got saddled with the presentation, yes?
For those who don’t know, Moeliker gave a really good TED talk. Worth watching. It’s not about suppressing uncomfortable information, though.
Next Wednesday, June 5th, is Dead Duck Day.
Makes you wonder what they are up to now.
The bug is called Leroy.
Upvoted. Then saw that that put the count at 422. So I had to downvote instead.
The article alleges, though without evidence, that the tracking is just an excuse to raise rates.
A quick search didn’t turn up quite the right statistics, but traffic fatalities have been seriously on the rise in the US. That probably implies higher payouts. (WP)
But also, when trackable unsafe drivers have to pay more (and trackable safe driver less), then the unsafe drivers will prefer to be untrackable. You may be on the receiving end of the recalculated actuary tables.
that will ultimately be used to create huge amounts of wealth for very few,
But… That is what these poisoning attacks are fighting for. They are attacking open image generators that can be used by anyone. You can use them for fun or for business, without having to pay rent to some owner who is not lifting a finger. What do you think will happen if you knock that out?
This attack doesn’t target Big Tech, at all. The model has to be open to pull off an attack like that.
This doesn’t have anything to do with tracking. This is supposed to sabotage free and open image generators (ie stable diffusion). It’s unlikely to do anything, though.
Hard to say what the makers want to achieve with this. Even if it did work, it would help artists just as much, as better DRM would help programmers. On its face, this is just about enforcing some ultra-capitalist ideology that wants information to be owned.
Text explaining why the neural network representation of common features (typically with weighted proportionality to their occurrence) does not meet the definition of a mathematical average. Does it not favor common response patterns?
Hmm. I’m not really sure why anyone would write such a text. There is no “weighted proportionality” (or pathways). Is this a common conception?
You don’t need it to be an average of the real world to be an average. I can calculate as many average values as I want from entirely fictional worlds. It’s still a type of model which favors what it sees often over what it sees rarely. That’s a form of probability embedded, corresponding to a form of average.
I guess you picked up on the fact that transformers output a probability distribution. I don’t think anyone calls those an average, though you could have an average distribution. Come to think of it, before you use that to pick the next token, you usually mess with it a little to make it more or less “creative”. That’s certainly no longer an average.
You can see a neural net as a kind of regression analysis. I don’t think I have ever heard someone calling that a kind of average, though. I’m also skeptical if you can see a transformer as a regression but I don’t know this stuff well enough. When you train on some data more often than on other data, that is not how you would do a regression. Certainly, once you start RLHF training, you have left regression territory for good.
The GPTisms might be because they are overrepresented in the finetuning data. It might also be from the RLHF and/or brought out by the system prompt.
I accidentally clicked reply, sorry.
B) you do know there’s a lot of different definitions of average, right?
I don’t think that any definition applies to this. But I’m no expert on averages. In any case, the training data is not representative of the internet or anything. It’s also not training equally on all data and not only on such text. What you get out is not representative of anything.
A) I’ve not yet seen evidence to the contrary
You should worry more about whether you have seen evidence that supports what you are saying. So, what kind of evidence do you want? A tutorial on coding neural nets? The math? Video or text?
That’s a) not how it works and b) not averaging.
Who exactly creates the image is not the only issue and maybe I gave it too much prominence. Another factor is that the use of copyrighted training data is still being negotiated/litigated in the US. It will help if they tread lightly.
My opinion is that it has to be legal on first amendment grounds, or more generally freedom of expression. Fair use (a US thing) derives from the 1st amendment, though not exclusively. If AI services can’t be used for creating protected speech, like parody, then this severely limits what the average person can express.
What worries me is that the major lawsuits involve Big Tech companies. They have an interest in far-reaching IP laws; just not quite far-reaching enough to cut off their R&D.
Why would averaging lead to repetition of stereotypes?
Anyway, it’s hard to say LLMs output what they do. GPTisms may have to do with the system prompt or they may result from the fine-tuning. Either way, they don’t seem very internet average to me.
You’re allowed to use copyrighted works for lots of reasons. EG satire parody, in which case you can legally publish it and make money.
The problem is that this precise situation is not legally clear. Are you using the service to make the image or is the service making the image on your request?
If the service is making the image and then sending it to you, then that may be a copyright violation.
If the user is making the image while using the service as a tool, it may still be a problem. Whether this turns into a copyright violation depends a lot on what the user/creator does with the image. If they misuse it, the service might be sued for contributory infringement.
Basically, they are playing it safe.
Yes, she said that. But what she said there just doesn’t make any sense.