kitonthenet
link
fedilink
51Y

Modular design

@crow@beehaw.org
link
fedilink
English
301Y

Also drowning “knowing how your code works”

can’t wait till chat gpt or its maintainers start injecting Skynet and other trojans

@Zeth0s@lemmy.world
link
fedilink
2
edit-2
1Y

I don’t see anything wrong in this picture. ML == funny!

“AI Casino” vs “old school nights banging heads on books”

Edit. Guys, it was a joke, do I really need to put /s?

Please, what is the established theory to create an n-1 dimensional space filling curve on the surface of an n-sphere?! I need to speed up my locality sensitive neural hash!

Any task that can be expressed as mostly translation is a good task to try with an LLM.

And you know what? Stakeholders tend to love LLMs, so have fun with your complicated problems while I build them by using the ancient technique of slapping some boilerplate together and combining it with the new ways of pasting error messages into chatgippity.

fmstrat
link
fedilink
English
91Y

Translation, but not categorization. Trying to get reliable, and more importantly, predictibly accurate, metadata from an LLM without serious training is a pain. ML algorithms are far better for this but certainly take more brainpower (in my experience so far).

TehPers
link
fedilink
English
31Y

There are a disproportionately large number of people who get one pretty demo and think LLMs are the solution to everything. Even for translations, I’d be interested to see how accurate the major models are in real world scenarios. We’ve been struggling hard to find any practical usage of LLMs that doesn’t require the user to be able to verify the output themselves.

@h3ndrik@feddit.de
link
fedilink
40
edit-2
1Y

I don’t think that was the point. The thing is, people replace calculators with that…

  • User: Assistant?
  • Assistant: * BEEP *
  • User: What is 21 divided by three?
  • Assistant: 52, my master.

Thing is, they only get some results right and hallucinate others. And you’re doing billions of matrix multiplications just to calculate 2+1.

Sure. You can go to a construction site with only your one favorite tool. And use it for everything. And it’s impressive to open a glass bottle of beer with a hammer and such. But I can guarantee you, you’ll be slower digging that hole than the guys using a proper tool like an excavator.

That’s not a translation problem, so LLMs are terrible for it.

Always use the right tool for the job. If there are a lot of nails to be hammered, you need a guy with a hammer.

Yeah, that was kind of my point. I think the meme picture means people throw it at everything. No matter what. And the next logical thing would be to strip the computer scientist out of the picture. We have Github Copilot now ;) Let AI decide if AI is the proper tool.

And: you don’t solve any fundamental problems if you don’t have the data for it. If the information isn’t in your data, the network will start guessing and it will be horrible.

Just chuck more data at it and hope for the best! It’s a pretty fun strategy even if it fails a lot

I’m convinced they’re just into machine learning because of the funny mistakes they make

That’s honestly one of the best parts about it. Reading error messages and thinking logically is boring compared to trying to guess how this stupid LLM got the idea to spit out what it did.

ML honestly sounds like a maddeningly dull profession to me because of this. It’s a cool technology but jiggling hyperparameters and then waiting would grate on me.

NX2
link
fedilink
261Y

Nvidia stock goes brrrr

Create a post

Post funny things about programming here! (Or just rant about your favourite programming language.)

Rules:

  • Posts must be relevant to programming, programmers, or computer science.
  • No NSFW content.
  • Jokes must be in good taste. No hate speech, bigotry, etc.
  • 1 user online
  • 61 users / day
  • 247 users / week
  • 417 users / month
  • 2.88K users / 6 months
  • 1 subscriber
  • 1.53K Posts
  • 33.9K Comments
  • Modlog