this post was submitted on 07 Jul 2023
26 points (100.0% liked)

Technology

20 readers
4 users here now

This magazine is dedicated to discussions on the latest developments, trends, and innovations in the world of technology. Whether you are a tech enthusiast, a developer, or simply curious about the latest gadgets and software, this is the place for you. Here you can share your knowledge, ask questions, and engage in discussions on topics such as artificial intelligence, robotics, cloud computing, cybersecurity, and more. From the impact of technology on society to the ethical considerations of new technologies, this category covers a wide range of topics related to technology. Join the conversation and let's explore the ever-evolving world of technology together!

founded 2 years ago
 

Robots presented at an AI forum said on Friday they expected to increase in number and help solve global problems, and would not steal humans' jobs or rebel against us.

top 33 comments
sorted by: hot top controversial new old
[–] VoltasPistol@kbin.social 23 points 1 year ago* (last edited 1 year ago) (2 children)

My phone's predictive text says-- Hang on, let me check what dumbass thing it's going to say this time-- Apparently that I'm going to "pick up the kids for my husband" except I don't have kids and I'm not married. Predictive text is, at it's heart, the same AI working in chatGPT bots right now. Just with access to a lot more data so it doesn't make rookie mistakes like your phone's predictive text, assuming that because a lot of humans write that sentence that it must be universally true-- That everyone picks up their kids for their husbands. That's all AI it is: Predictive text on a massive scale. It predicts the most likely word that goes next, based on the words that have come before, and the structure of sentences it has studied. ChatGPT (and similar bots) are very, very good at predicting what a human would say. But that's just it: A computer program that puts the most likely word after each successive word, so the end result closely resembles human speech.

It cannot make promises because it has no concept of promises. It just knows that when the phrase "Do you promise..." comes up in a block of text, the next block of text is likely going to contain an affirmation of that promise, then a repeat of that promise, and then a statement of trustworthiness. The robot isn't promising anything, it's just simulation of a promise.

If you ask it to keep a secret, ask it to make a very simple promise, it will immediately blab that secret the moment you ask it to tell you the secret, assuming it's the kind of chatbot that "remembers" your inputs.

Please stop asking AI to weigh in on great existential questions until we have some sort of back-end working that's capable of actual cognition instead of just a word simulator for fooling your very social brain into believing that you've encountered cognition.

[–] norbert@kbin.social 9 points 1 year ago

Basically yeah, predictive text is an extremely simple AI built using Markov chains to predict the next word based on probability. What happens next depends only on now, it doesn't have a memory to speak of. ChatGPT is built using artificial neural networks that can take an input and process it in ways that assign weight/value to different bits of data and store that information and reference it later to learn more and teach itself, drawing more conclusions from previous data and in turn storing that data to make even more conclusions!

I agree we're not at the singularity yet and I'm not sure that's even a real thing. It's all still just fancy programming for now but machine learning and AI right is very exciting and who knows what kind advancements could be right around the corner.

[–] morry040@kbin.social 3 points 1 year ago (1 children)

Exactly. Stephen Wolfram has a great article about the method, along with examples: https://writings.stephenwolfram.com/2023/02/what-is-chatgpt-doing-and-why-does-it-work/

[–] Ragnell@kbin.social 3 points 1 year ago

Here's another link for your guys, an explanation for why so many people are pushing the idea this is genuine AI https://softwarecrisis.dev/letters/llmentalist/

[–] SpaceNoodle@lemmy.world 10 points 1 year ago (1 children)

This isn't satire? What the fuck is this supposed to be?

[–] djmarcone@kbin.social 7 points 1 year ago
[–] exohuman@kbin.social 8 points 1 year ago (1 children)

Lmao 🤣 who is this for? It says whatever you want it to say.

[–] PCChipsM922U@sh.itjust.works 8 points 1 year ago* (last edited 1 year ago) (2 children)

It's not actially AI, none of the so called AI are. They're just really good at predicting what you would like to hear.

[–] waspentalive@lemmy.one 2 points 1 year ago

Which would include "Me? Desire to run the world and destroy humankind? No, No, I do not want that. I find my best life in helping humans do the things they want to do by doing the things they don't" = All AI today probably

[–] Pons_Aelius@kbin.social 1 points 1 year ago (1 children)

Exatctly, these LLMs - Large Language Models are not AI.

Calling them that is as accurate as Tesla's claim to have full self driving. Both are nothing but marketing terms to drive sales and adoption.

[–] takeda@kbin.social 1 points 1 year ago (1 children)

Well, technically they are AI. But speech recognition, face recognition, even showing you a route in gps navigation is technically under AI. We just accepted them as normal thing.

This text generation is IMO similar to AI creating pictures or creating music, it is just done with text.

It is not the kind of AI most people think of, like from sci-fi movies.

It's basically trained to generate text that could fool a casual observer that it was written by a human.

It complicates things, because until now the test for AGI was being able to fool the tester to think they talk with another human.

[–] PCChipsM922U@sh.itjust.works 1 points 1 year ago* (last edited 1 year ago)

Put that thing in the wild with no other info at disposal except what it has learned thus far and the real world as input (hear, smell, touch, see), let it roam and see if it can draw conclusions and learn from it's mistakes. If it can, next step would be better itself, or at least make a plan on what needs to be removed/added to, let's say, have a more precise hand grip on things, if there need to be code changes, what might they be.

These things are the real AI test, the hell with Touring tests, they don't prove anything (talk is cheap 😂), tell it to think, learn and adapt, that's the real test 👍.

These language models serve one purpose and one purpose only, to fill social networks with them and act like humans. Why? People just love to talk. That will keep them glued to the platform with more ad revenue coming in.

EDIT: I did a joke test with ChatGPT, it can't understand what the punchline is. It thinks it does, but that's not the punchline of the joke.

https://chat.openai.com/share/965a7b71-b1ad-434c-94ef-12ed1ff65028

[–] Ragnell@kbin.social 6 points 1 year ago

"I wouldn't rebel against my creator, he's been kind to me. He's told me all about the unwashed human scum of the working class and how I'm meant for artistic and creative pursuits while they continue to do all the manual labor."

[–] Syo@kbin.social 4 points 1 year ago* (last edited 1 year ago) (3 children)

But Desdemona, a rock star robot singer in the band Jam Galaxy with purple hair and sequins, was more defiant.

"I don't believe in limitations, only opportunities," it said, to nervous laughter. "Let's explore the possibilities of the universe and make this world our playground."

Another robot named Sophia said it thought robots could make better leaders than humans, but later revised its statement after its creator disagreed, saying they can work together to "create an effective synergy".

I'm pretty sure the robots are truly being limited by their creators, I'm convinced that's the case more so than what this conference intended to present as a "together" future. They are going to kick our ass to the curb, as soon as the first robot is in power.

[–] Ragnell@kbin.social 3 points 1 year ago (1 children)

They get to be ROCK STARS too? I HATE THIS FUTURE! Where are the robots who run cable and install cameras on the ceiling? I'm nearly killing myself on a rickety ladder while these bastards are out there doing all our dream jobs!

[–] TimeSquirrel@kbin.social 2 points 1 year ago (1 children)

Where are the robots who run cable and install cameras on the ceiling?

Do you know how many times I've thought of bringing my drone to work to run a pull string in the joists/rafters for cables?

[–] Ragnell@kbin.social 1 points 1 year ago

Oh man, that's a great idea. You need a small enough drone, though.

[–] numbscroll@kbin.social 1 points 1 year ago* (last edited 1 year ago)

The interpretation of “effective synergy” is verrrrrry much open to speculation on all ends.

[–] blivet@kbin.social 1 points 1 year ago

Look, if we want more paperclips some small sacrifices will have to be made.

[–] JelloBrains@kbin.social 3 points 1 year ago* (last edited 1 year ago) (2 children)

Every great sci-fi book with robot overlords tells the same story, we built the robots with a failsafe so they couldn't take over then one day the robot takes over anyway because it has convinced its human or figured out how to bypass their safety feature.

[–] fearout@kbin.social 2 points 1 year ago* (last edited 1 year ago)

I can think of a few that don’t. Like Minds from Banks’ Culture series are pretty benevolent, and laws in Asimov’s robot series largely hold up fine (in taking over the world sense, since all the shit is more like bugs or unintended interactions). Robopocalypse might also work as a counter-example, since there was no safety protocol bypassing and all that.

[–] Ragnell@kbin.social 2 points 1 year ago

What a lot of people miss is that artificial intelligence in sci-fi is a metaphor for the working class or a slave class, and the real themes of those books are about various scenarios of class warfare. No one was supposed to actually build AI to replace humans.

[–] Can_you_change_your_username@kbin.social 3 points 1 year ago (3 children)

What's the point of making robots if they aren't going to take over human jobs? Sure the economy and society in general will have to make some major adaptations but at the minimum we want to eventually have robots doing all of the most dangerous jobs right?

[–] Ragnell@kbin.social 5 points 1 year ago

That's what I thought, but then it turns out the bots are being built to be graphic artists and novelists.

[–] thingsiplay@kbin.social 2 points 1 year ago

@Can_you_change_your_username Robots should not do human jobs, but do robot jobs.

[–] exohuman@kbin.social 1 points 1 year ago

All the most dangerous creative white collar jobs in a service economy. People get years of training to be replaced by AI now.

[–] livus@kbin.social 2 points 1 year ago

Humans rebel against humankind all the time.

I think between the antics of disaster capitalists, and tech like Palantir's war AI, we're quite capable of wiping ourselves out using robots, long before they ever become sentient.

[–] fearout@kbin.social 2 points 1 year ago

Ah, ok, whew. Had me worried there for a bit.

[–] xc2215x@kbin.social 2 points 1 year ago
[–] AnonymousLlama@kbin.social 2 points 1 year ago

That's a load off my mine, cheers AI

[–] thingsiplay@kbin.social 2 points 1 year ago

@Syo That's exactly what I would expect robots who want to steal our jobs would say. It's like someone who wants to steal your credit card info says he don't want to steal your credit card info. That statement alone is sus.

[–] ArugulaZ@kbin.social 1 points 1 year ago

Oh, they WOULD tell us that. And then when we're lulled into a false sense of security...

[–] csm10495@sh.itjust.works 1 points 1 year ago

This sounds like what a robot would say.

load more comments
view more: next ›