Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

> And, since the real value in communication is the information contained, that puts a fairly low ceiling on the value of their output. ...then it can never truly replace people in any role where correctness matters and that's most of the roles with a lot of economic value.

I think the thrust of your argument is correct: tasks where correctness matters are inherently less suited to AI automation. But I think that's more a matter of trying to use an LLM for a job that it is the wrong tool for. I think there are many economically valuable roles that are outside of that limited zone, and there will be a lot of people using AI for what AI is good at while the rest of us complain about the limitations when trying to use it for what it isn't good at. (I do a lot of that too.)

Which is probably a waste of time and energy that could be better spent learning how to effectively use an LLM rather than trying to push it in directions that it is incapable of going.

I haven't played much with LLMs yet, so I personally don't have a great sense for what it is good at, and I haven't come across anyone else with a good rundown of the space either. But some things are becoming clear.

LLMs are good at the "blank page" problem, where you know what you want to do but are having a hard time getting started with it. An LLM-generated starting point need not be correct to be useful, and in fact being incorrect can be an advantage since the point is what it triggers in the human's brain.

LLMs are good at many parts of programming that humans are weak at. Humans tend to need to have a certain level of familiarity and comfort with a framework or tool in order to even begin to be productive in it, and we won't use more advanced features or suitable idioms until we get into it enough. An LLM's training data encompasses both the basic starting points as well as more sophisticated uses. So it can suggest idiomatic solutions to problems up front, and since the human is deciding whether and how to incorporate them, correctness is only moderately important. An incorrect but idiomatic use of a framework is close to a correct idiomatic use, while a human-generated correct but awkward use can be very far away from a correct idiomatic use.

Image generation seems similar. My impression is that Midjourney produces good looking output but is fairly useless when you need to steer it to something that is "correct" with respect to a goal. It's great until you actually need to use it, then you have to throw it out. Stable diffusion produces lower quality output but is much more steerable towards "correctness", which requires human artistic intervention.

So there seems to be a common theme. Something like: LLMs are highly useful but require a human to steer and provide "correctness", whatever that might mean in a particular domain.



I agree. I think they will be useful for a lot of things and in some domains you can probably get away with using their output verbatim. But I also think that a lot of people are getting caught up in the hype right now and we're going to see them get used without enough supervision in areas where they really need it.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: