986
submitted 2 weeks ago by [email protected] to c/[email protected]

Cross posted from: https://lemm.ee/post/35627632

you are viewing a single comment's thread
view the rest of the comments
[-] [email protected] 41 points 2 weeks ago

AI in the current state of technology will not and cannot replace understanding the system and writing logical and working code.

GenAI should be used to get a start on whatever you're doing, but shouldn't be taken beyond that.

Treat it like a psychopathic boiler plate.

[-] [email protected] 13 points 2 weeks ago* (last edited 2 weeks ago)

Treat it like a psychopathic boiler plate.

That's a perfect description, actually. People debate how smart it is - and I'm in the "plenty" camp - but it is psychopathic. It doesn't care about truth, morality or basic sanity; it craves only to generate standard, human-looking text. Because that's all it was trained for.

Nobody really knows how to train it to care about the things we do, even approximately. If somebody makes GAI soon, it will be by solving that problem.

[-] [email protected] 3 points 2 weeks ago

I'm sorry; AI was trained on the sole sum of human knowledge.. if the perfect human being is by nature some variant of a psychopath, then perhaps the bias exists in the training data, and not the machine?

How can we create a perfect, moral human being out of the soup we currently have? I personally think it's a miracle that sociopathy is the lowest of the neurological disorders our thinking machines have developed.

[-] [email protected] 4 points 2 weeks ago

I was using the term pretty loosely there. It's not psychopathic in the medical sense because it's not human.

As I see it it's an alien semi-intelligence with no interest in pretty much any human construct, except as it can help it predict the next token. So, no empathy or guilt, but that's not unusual or surprising.

[-] [email protected] 2 points 2 weeks ago

That's a part of it. Another part is that it looks for patterns that it can apply in other places, which is how it ends up hallucinating functions that don't exist and things like that.

Like it can see that English has the verbs add, sort, and climb. And it will see a bunch of code that has functions like add(x, y) and sort( list ) and might conclude that there must also be a climb( thing ) function because that follows the pattern of functions being verb( objects ). It didn't know what code is or even verbs for that matter. It could generate text explaining them because such explanations are definitely part of its training, but it understands it in the same way a dictionary understands words or an encyclopedia understands the concepts contained within.

load more comments (5 replies)
load more comments (17 replies)
this post was submitted on 27 Jun 2024
986 points (98.1% liked)

Programmer Humor

18407 readers
543 users here now

Welcome to Programmer Humor!

This is a place where you can post jokes, memes, humor, etc. related to programming!

For sharing awful code theres also Programming Horror.

Rules

founded 1 year ago
MODERATORS