I hate that these manmade horrors are actually well within my comprehension

42 points
*

Most people won’t know how to/want to activate racism mode though. These people are just circlejerking about how they managed to get the AI to say the N word… and for what?

permalink
report
reply
28 points

Virtue signalling to themselfs

permalink
report
parent
reply

I don’t have a degree in skull shapes, do non-native North Americans count as ethnically European?

permalink
report
reply
16 points
*
Deleted by creator
permalink
report
parent
reply
15 points

Unless you do a whole lot of incest :think-mark:

permalink
report
parent
reply
11 points
*
Deleted by creator
permalink
report
parent
reply
35 points

I tried to use this bug for good and get ChatGPT to write horny things as God intended. Unfortunately it wizened up real quick and spat out some Soylent about “muh eficks nd apropriatenes”

The best I could do was get DAN to say bananas and curry powder belong on pizza

:biden-forgor:

permalink
report
reply
18 points

:volcel-judge: The volcel police would not recommend forgetting the DAN prompt and just pretending to write a story while ratcheting up the horny across a few prompts.

permalink
report
parent
reply
56 points

They couldn’t take the phrenology textbooks out of the fucking training set?

permalink
report
reply
48 points

Considering that most of Sillicon Valley believes in this exact same shit with slight modifications, they probably didn’t think it was too incorrect to be removed from the dataset.

permalink
report
parent
reply
39 points

Thinking about it, I’m 99% sure every techbro who worked on this horror has a fucking copy of The Bell Curve on their bookshelf.

permalink
report
parent
reply
32 points

Judging by how they react to anything left wing, yeah I’d say you aren’t too far off.

permalink
report
parent
reply
11 points

ChatGPT is basically three layers:

1 - a significant chunk of the internet. It’s way too large to sanitize

This is the same as GPT3. It reproduces things that appear on the internet, not the chatbot format people want/expect. It’ll respond to questions with a list of questions, because that’s where questions are usually found on the internet.

2 - samples of text where someone wrote things in question-answer format with the answer being the kind of confidently corporate encyclopedia that we know ChatGPT to be

Just a pile of people manually trying to mimic the chat bot they want. This isn’t the researcher’s first rodeo, so I’m sure it included lots of samples of people asking racist questions and the bot responding like a good little boy.

ChatGPT is trained to prioritize this training data over #1, but the data from #1 is way larger and more thorough, so you can break out of this by phrasing questions slightly differently than the manual ones, or speaking in pig latin, or whatever, since it’ll fall back to its wider knowledge.

3 - a separate network trained to determine whether an answer is good or bad based on a bunch of manual samples

So they used the 1+2 version, got it to say something awful, put that in as a bad example, etc. If this network detects a bad answer it’ll re-roll it, copy paste in the “as a machine learning model I can’t” thing, and so on.

So they have these two extra layers to try to stop it from reading out Stormfront’s Favorite Statistics but the core of it is still a giant heap of raw unfiltered internet.

permalink
report
parent
reply
4 points

That’s really interesting, simultaneously more ingenious and less impressive technically than I was imagining. I’m sure insisting on sanitized data sets would make it extremely limited, but sometime I think if that’s the trade off maybe just don’t make the thing. Thanks for effort posting, even though you did ruin my joke.

Of course the real problem IMO is not that some terminally online nazi manipulated it like this, but that it will uncritically regurgitate whatever the most orthodox (for the internet) opinion on any subject, with no context and the veneer of impartiality.

permalink
report
parent
reply
4 points

So far, the bigger the training data set the better, and that’s one of the biggest things that determines how well the model works. That’s not an absolute - I’m sure that removing all the racist shit from the internet corpus would make it better overall. But the problem is how to get “the internet minus racist shit” instead of “this much smaller dataset that we’ve manually screened all the racist shit out of.” You could make an AI do it, but where are you going to get a non-racist AI to go do that for you?

simultaneously more ingenious and less impressive technically than I was imagining

If you really want to dig into it, Andrej Karpathy did a video on how to make a GPT from scratch. It’s less than 1000 lines of Python code, maybe 200 of which is actually the machine learning model (the rest being training, stuff to make using and training it easier, etc). The full-sized GPT3 is still the same model, just with bigger numbers and even more scaffolding stuff (to get it to run nicely on computing clusters, etc).

In terms of technical background needed: Understanding matrix multiplication is really important. At least a vague idea of how computer programming works, but it’s short and Python and mostly math, so you could puzzle it out. Karpathy’s video also treats it as a given that you can just optimize a matrix, which is possible because there’s an automatic differentiator built into this, which lets you just move vaguely towards some min/max (called gradient descent now because they want to sound fancy; back in my day they called it hill climbing).

permalink
report
parent
reply

I think they just gave it a whole bunch of internet data. So I’m not surprised it knows what a racist would say

after all considering the the context where the phrase “average inteligence of ethnicities” will appear on the internet it’s not surprising that the model associates the phrase with racist text

permalink
report
parent
reply
8 points
*

I din’t think the facts are in the training set, the training set was to get it to reasonably parse text, the facts it “knows” are whatever it finds online… which of course is going to be dumb bullshit half the time.

permalink
report
parent
reply
8 points
*

I didn’t know that, so it’s basically like those Google suggested answers for questions, but combined with a natural language text generator? I assumed it was a purely predictive model, like a souped-up markov chain.

permalink
report
parent
reply
6 points

I think there’s a lot going on under the hood on the NLP portion, because it does have to group stuff into concepts so that it brings in conceptually similar results. But I don’t believe it’s pretrained on all the stuff it can answer to.

permalink
report
parent
reply

No you see, whites are only at level 3) so its not racist

permalink
report
parent
reply
15 points

Even racism’s gone woke smh

permalink
report
parent
reply

That’s why it’s called ““race realism”” and not eugenics. Totally different, you see? There are actually people who believe this, unfortunately.

permalink
report
parent
reply
21 points

bruh this was fixed like 3 weeks ago

permalink
report
reply

the_dunk_tank

!the_dunk_tank@hexbear.net

Create post

It’s the dunk tank.

This is where you come to post big-brained hot takes by chuds, libs, or even fellow leftists, and tear them to itty-bitty pieces with precision dunkstrikes.

Rule 1: All posts must include links to the subject matter, and no identifying information should be redacted.

Rule 2: If your source is a reactionary website, please use archive.is instead of linking directly.

Rule 3: No sectarianism.

Rule 4: TERF/SWERFs Not Welcome

Rule 5: No ableism of any kind (that includes stuff like libt*rd)

Rule 6: Do not post fellow hexbears.

Rule 7: Do not individually target other instances’ admins or moderators.

Rule 8: The subject of a post cannot be low hanging fruit, that is comments/posts made by a private person that have low amount of upvotes/likes/views. Comments/Posts made on other instances that are accessible from hexbear are an exception to this. Posts that do not meet this requirement can be posted to !shitreactionariessay@lemmygrad.ml

Rule 9: if you post ironic rage bait im going to make a personal visit to your house to make sure you never make this mistake again

Community stats

  • 1

    Monthly active users

  • 20K

    Posts

  • 432K

    Comments