this post was submitted on 05 Jan 2024
137 points (95.4% liked)

Technology

57435 readers
3218 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 3 points 7 months ago

I agree completely. I think AI can be a valuable tool if you use it correctly, but it requires you to be able to prompt it properly and to be able to use its output in the right way - and knowing what it's good at and what it's not. Like you said, for things like brainstorming or looking for inspiration, it's great. And while its artistic output is very derivative - both because it's literally derived from all the art it's been trained on and simply because there's enough other AI art out there that it doesn't really have a unique "voice" most of the time - you could easily use it as a foundation to create your own art.

To expand on my asking it questions: the kind of questions I find it useful for are ones like "what are some reasons why people may do x?" or "what are some of the differences between y and z?". Or an actual question I asked ChatGPT a couple of months ago based on a conversation I'd been having with a few people: "what is an example of a font I could use that looks somewhat professional but that would make readers feel slightly uncomfortable?" (After a little back and forth, it ended up suggesting a perfect font.)

Basically, it's good for divergent questions, evaluative questions, inferent questions, etc. - open-ended questions - where you can either use its response to simulate asking a variety of people (or to save yourself from looking through old AskReddit and Quora posts...) or just to give you different ideas to consider, and it's good for suggestions. And then, of course, you decide which answers are useful/appropriate. I definitely wouldn't take anything "factual" it says as correct, although it can be good for giving you additional things to look into.

As for writing code: I've only used it for simple-ish scripts so far. I can't write code, but I'm just about knowledgeable enough to read code to see what it's doing, and I can make my own basic edits. I'm perfectly okay at following the logic of most code, it's just that I don't know the syntax. So I'm able to explain to ChatGPT exactly what I want my code to do, how it should work, etc, and it can write it for me. I've had some issues, but I've (so far) always been able to troubleshoot and eventually find a solution to them. I'm aware that if want to do anything more complex then I'll need to expand my coding knowledge, though! But so far, I've been able to use it to write scripts that are already beyond my own personal coding capabilities which I think is impressive.

I generally see LLMs as similar to predictive text or Google searches, in that they're a tool where the user needs to:

  1. have an idea of the output they want
  2. know what to input in order to reach that output (or something close to that output)
  3. know how to use or adapt the LLM's output

And just like how people having access to predictive text or Google doesn't make everyone's spelling/grammar/punctuation/sentence structure perfect or make everyone really knowledgeable, AIs/LLMs aren't going to magically make everyone good at everything either. But if people use them correctly, they can absolutely enhance that person's own output (be it their productivity, their creativity, their presentation or something else).