Saved by sari and
All the Hard Stuff Nobody Talks About When Building Products With LLMs
People need to be more thoughtful building products on top of LLMs. The fact that they generate text is not the point.
Linus Leestream.thesephist.comReply
reply
LinuxSpinach
•
5h ago
^ this. And especially classification as a task, because businesses don’t want to pay llm buck... See more
r/MachineLearning - Reddit
Nicolay Gerold added
We're doing NER on hundreds of millions of documents in a specialised niche. LLMs are terrible for this. Slow, expensive and horrifyingly inaccurate. Even with agents, pydantic parsing and the like. Supervised methods are the way to go. Hell, I'd take an old school rule based approach over LLMs for this.
The biggest thing any ML practitioner realizes when they step out of a research setting is that for most tasks accuracy has ... See more
Ask HN: What are some actual use cases of AI Agents right now? | Hacker News
Nicolay Gerold added
You are assuming that the probability of failure is independent, which couldn't be further from the truth. If a digit recogniser can recognise one of your "hard" handwritten digits, such as a 4 or a 9, it will likely be able to recognise all of them.
The same happens with AI agents. They are not good at some tasks, but really really food at others.
Avital Balwit • My Last Five Years of Work
Max Beauroyre added
"A key challenge of (LLMs) is that they do not come with a manual! They come with a “Twitter influencer manual” instead, where lots of people online loudly boast about the things they can do with a very low accuracy rate, which is really frustrating..."
Simon Willison, attempting to explain LLM
Johann Van Tonder added
Ask HN: What are some actual use cases of AI Agents right now? | Hacker News
Nicolay Gerold added
Is this a good thing or a bad thing? I’m not sure.
A great example of this is frontend... See more
Shortwave — rajhesh.panchanadhan@gmail.com [Gmail alternative]
Nicolay Gerold added