Hacker Newsnew | past | comments | ask | show | jobs | submit | 3836293648's commentslogin

There is discussion about this in the Rust world, though no attempts at implementation (and yet further from stabilisation)

Maybe if you owned the tree, not if someone a few houses down does

This is in response to all the pushback they got from that

But this also means tiny context windows. You can't fit gpt-oss:20b + more than a tiny file + instructions into 24GB

Gpt-oss is natively 4-bit, so you kinda can

You can fit the weights + a tiny context window into 24GB, absolutely. But you can't fit anything of any reasonable size. Or Ollama's implementation is broken, but it needs to be restricted beyond usability for it not to freeze up the entire machine when I last tried to use it.

LLMs do typically encode a confidence level in their embeddings, they just never use it when asked. There were multiple papers on this a few years back and they got reasonable results out of it. I think it was in the GPT3.5 era though


It's mostly AI slop, but they did exist before AI (and they were miserable back then too)


Are you seriously comparing discrimination based on factors noone can control to a group literally defined by a choice they made? And you think that's a good faith argument?


People hate C because it's hard, people hate C++ because it truly is rubbish. Rubbish that deserved to be tried but that we've now learned was a mistake and should move on from.


Because llm tokens don't map cleanly to what the compiler sees as a token. If coding is all LLMs will be good for this will surely change


What on Earth have you used to get reasonable results out of a local model?

I've tried at every new model release (that can run on my 24GB card) and everything is still entirely useless.

I'm not writing web stuff though.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: