AI couldn’t do this a year ago, it required computer hardware that was supercomputer levels of expensive to even create something like this. IMO development was actually held back by crypto and covid-19. Now AI is the #1 focus of the techbros and it isn’t going to slow down. This shit is going to put so many journalists, artists, and even programmers out of work. I don’t know how else to explain this, HUMANITY LITERALLY CREATED AI THIS YEAR. WE MADE FUCKING SKYNET!

You want to talk about technological progress, this shit mogs fusion, it mogs the vaccines, it mogs whatever dumb space colonization shit we did. We made fucking AI! I bet we will have sentient AI in our lifetime. And what are we going to do with this stuff? Porn, lots of porn. Deepfakes of celebrities and politicians sacrificing children to moloch, dead actors staring in new movies, a new album by Tupac, fake war footage, fake everything.

Have you ever heard about how a monkey can write Shakespeare given enough time? We have fucking done that, we pressed random buttons enough times that we ended up with something legible. Now we turn memes into real people.

We need a butlarian jihad or some shit.

this isn’t close to sentience technical image processing to look like a person is not nearly as difficult as thought or having ideas.

It might lead to more easily fakeable news but the news is already inundated with lies and nonsense

permalink
report
reply
26 points

Yeah, and while AI couldn’t do this specific thing until this year, this isn’t an AI revolution so much as small improvements over what it has been able to do for 10-20 years. “Make an AI remake an image as if it’s a van Gogh” has been in deep learning courses for years now, and I don’t see that being significantly different from this. As for putting artists out of work, I’m not convinced movies and video game studios are going to ship stuff done by AI, probably AI generated assets will just get incorporated into stuff like photoshop as a base for artists to work from.

permalink
report
parent
reply

It might be used a bit like CGI but studios also might well respond to a new source of effects by just including even more effects

permalink
report
parent
reply
13 points

Hopefully it just cuts down hours of artists, a bit of automation where people get worked extreme hours could be a net good for the workforce. They’ll probably need to organize and fight for it still.

permalink
report
parent
reply
10 points
*
Deleted by creator
permalink
report
parent
reply

I’m thinking of those tabloid papers with the photoshopped covers that used to be all over grocery store shelves.

permalink
report
parent
reply

also we shouldn’t really worry at all about sentient AI so much as sapient AI. Computers sense electrical voltages to function. Computers have been able to see and hear for decades.

fucken scifi writers fucking up terminology a hundred years ago.

permalink
report
parent
reply
38 points

I bet we will have sentient AI in our lifetime.

You will lose that bet unless the immortality losers figure something out. The “AI” we have is no more sentient than a list of numbers on a page, and will continue to not be sentient regardless of how big the page is. Unless you take the, in my opinion, very dim view that sentient beings are still just stimulus response machines, in which case your lightswitch is sentient as well.

permalink
report
reply
5 points
*

Machine learning algorithms will be sentient in the same way capital is. A neural network functions almost identically to human social systems. Where the data is the historical initial conditions, the trained “brain” is the mode of production, and the output is the productive sum of the historical conditions and the instantaneous social order.

You could say that the whole of society is intelligent, but it’s a different kind of intelligence to individual intelligence. The historical data that these systems (both society and machine learning) are trained on are built on the outputs of vast amount of human systems while the human system is built on a vast amount of interactions with the material world.

Machine learning is incapable of (at least now) replicating the human experience because it’s nerve endings don’t understand heat, cold, pain, light, etc, they understand written language. A system designed to form a social being that’s above all humans but totally separate from them.

This dynamic is also why machine learning will be absolutely crucial to any and all centrally planned economies in the future. No human is able to process the amount of raw data that a specialized machine can, and no machine can process the vast amount of sensory inputs a human can. And as these machine systems begin to get closer to the levers of social order, the conditions for revolution will begin appearing more frequently and aggressively. Suddenly the social god has buttons you can press and code you can modify, and the working class has 200 years of experience pressing buttons.

permalink
report
parent
reply
4 points

You could say that the whole of society is intelligent, but it’s a different kind of intelligence to individual intelligence.

i think that using the same word for those concepts is probably suboptimal

permalink
report
parent
reply
3 points
*

It is, but that was kinda my point that “Artificial Intelligence” as we have it now is a control system that functions almost identically to societal modes of production. So the absurdity of calling the whole of society “Intelligent” in the same way we call humans intelligent is nonsense.

Machine learning is a control loop that’s optimized to return an output for a series of inputs, in the most common case, text strings to text strings or text strings to pixel values. Capitalist society is a control loop that takes labor and resources and outputs profit.

Humans are more complex as the inputs and outputs of our biological systems are essentially infinitely less binary. Hypothetically you could use machine learning systems to simulate every nerve ending and biological state of the human body, but as of now our best machines can’t even do that for a single cell.

permalink
report
parent
reply
2 points

extelligence

as in, external. the inverse of to intelligence

permalink
report
parent
reply
4 points
*

very dim view that sentient beings are still just stimulus response machines

is there a materialist alternative to this view? Sentience to me seems to be an emergent property of a very complicated biological machinery, and if that machinery becomes damaged, so does the sentience. But I’m open to hearing your thoughts on the matter.

permalink
report
parent
reply
3 points

I mean, materially there probably isn’t even free will, there’s no way to make that suddenly pop out of particles. I just don’t bother factoring that into a worldview that actually affects the lives of other people, for the same reason I discard what could very well be biological fact that we are all stimulus response machines. And in fact I agree with you, treating free will and sentience and any other hard to pin down concept as an emergent property is probably the right way to do it - it’s like a phase transition.

My issue with applying those things to any kind of AI running on a von Neumann architecture is really just that I’m not convinced that system, which regardless of what AI nerds will tell you is still a billion times less complex than a nervous system, can ever hit that phase transition. Whether that’s a difference in kind or just in scale, I can honestly say I have no idea.

But more than anything, I think we as a species are so desperate for something like us, we’re at a real risk of anthropomorphizing what is 100% just a function spitting out what it was optimized for.

No idea if any of the above is particularly coherent or even a good argument, and of course this is putting aside all the actual concerns with the AI we have, re: intellectual property and accuracy and labor and everything else. Thanks for reading either way

permalink
report
parent
reply
2 points

But more than anything, I think we as a species are so desperate for something like us, we’re at a real risk of anthropomorphizing what is 100% just a function spitting out what it was optimized for.

There was a poll a while back that asked people what they thought of and expected from artificial intelligence. Apparently, the majority of people said it would be nice to finally have someone to talk to.

permalink
report
parent
reply
1 point

Thanks for sharing your thoughts.

My issue with applying those things to any kind of AI running on a von Neumann architecture is really just that I’m not convinced that system, which regardless of what AI nerds will tell you is still a billion times less complex than a nervous system, can ever hit that phase transition.

not for a while, certainly, but I think what we have created is already so much more complex than what we would have imagined ourselves to be capable of only a few centuries ago.

But more than anything, I think we as a species are so desperate for something like us, we’re at a real risk of anthropomorphizing what is 100% just a function spitting out what it was optimized for.

I think we’ve already done this with ourselves, and that it’s not necessarily a bad thing. Humanization is the opposite of dehumanization. I think sufficiently complex material phenomena that show signs of self awareness, pleasure, pain, empathy, and so on, deserve to be treated with respect and dignity, and not mistreated or used like an object. Since animals and humans deserve that, so would any emergent artificial intelligence (which I still think is a very very long way off, but possible through several avenues)

permalink
report
parent
reply
2 points

there’s the term sapience as in homo sapiens, and a bunch of non-human animals for short of that as well.

some SF writers used sentient life to refer to aliens a long time ago and people just rolled with the term even though they usually mean something more specific than what plants and termites have going on internally.

permalink
report
parent
reply

Hexbear stop becoming doomers over Latest Tech Fad #37 challenge (impossible!)

AI can only make art by stealing existing art and remixing it, which is legally questionable (in a “Disney is going to pull out the lawyers and lobbyists” sort of way) and also still requires people. News has been written by algorithms for a decade now anyway. Programmers don’t just write arbitrary blackbox functions for a living, you need a person to actually architect a system and figure out how data flows through it and perform meaningful actions which AI cannot do.

AI will not be sentient in anyone reading this’s lifetime. This is all a jumped up chatbot. Please stop turning into the computer equivalent of antivaxxers because someone posted an article about someone’s art getting stolen by a grifter AI, it’s not that big of a deal.

permalink
report
reply

News has been written by algorithms for a decade now anyway

as a pro tip if you write an article and give it to a reporter a lot of them will barely even read it before publishing it

permalink
report
parent
reply
2 points

Hexbear stop becoming doomers over Latest Tech Fad #37 challenge (impossible!)

:yea:

permalink
report
parent
reply
23 points

Nah, we didn’t create this in a year. People have worked on this since the 1960s. They got significantly better around 2014 when people switched to using GANs, and the widespread release of latent diffusion models released in the past year (that we’ve also been working on since 2015 or so) was another huge change. They also aren’t close to being AI in the sci-fi sense, AI is basically just a marketing term in this case and it’s probably more accurate or less misleading to just call it machine learning or deep learning.

It’s likely that people are overestimating how much this will put people out of work. These things can’t really create anything novel without a lot of manual guidance, which is a fundamental limitation as of now and it’s not certain when that may be overcome. Mostly it combines concepts well, and as you get more specific it starts to make more and more mistakes. Text gets incoherent and inaccurate, images with too much in the prompt end up very distorted if they even follow all of the prompt at all, code that’s more complicated than what you might find in documentation samples is unlikely to run.

There is quite a lot it can do to make someone already in one of these positions able to do more. AI upscalers are amazing, people have been using them for years now to restore old videos with great results (they were also trained off of unlicensed works and yet nobody complains about that). You can also get a lot more out of AI image generators if you actually have some amount of art skills – if you know how to compose a scene or draw hands better than the AI, you can sketch those things out and let it fill in the details, so you can make up for its deficiencies. As far as text models go, I don’t think we’re at the same point of progression for those as we are for image models, and I would think we may need a breakthrough on the level of what latent diffusion did to supersede GANs. GPT-3 is pretty great compared to other text models but its hardware requirements are astronomical. Half the reason that AI image generation is so widespread right now is the fact that you can run or train it on widely available consumer hardware, meanwhile the hardware to run GPT-3 (if the model weights were even public) costs as much as a luxury car. You probably won’t see anything too flashy until (and unless) that is solved.

permalink
report
reply
22 points
*

We need a butlarian jihad or some shit.

"The enormous destruction of machinery […] known as the Luddite movement, gave the anti-Jacobin governments […] a pretext for the most reactionary and forcible measures. It took both time and experience before the workpeople learnt to distinguish between machinery and its employment by capital, and to direct their attacks, not against the material instruments of production, but against the mode in which they are used. :marx:

(marx said butlerian jihad is canceled, sorry i don’t make the rules)

permalink
report
reply
8 points

we really do be just rehashing everything marx said but in meme format :monke-beepboop:

permalink
report
parent
reply
2 points

based

permalink
report
parent
reply